Efficient Subgraph Matching by Postponing Cartesian Products. Matthias Barkowsky

Size: px
Start display at page:

Download "Efficient Subgraph Matching by Postponing Cartesian Products. Matthias Barkowsky"

Transcription

1 Efficient Subgraph Matching by Postponing Cartesian Products Matthias Barkowsky 5

2 Subgraph Isomorphism - Approaches search problem: find all embeddings of a query graph in a data graph NP-hard, but can often be solved efficiently in practice general approach: iteratively map query vertices to data vertices -> ordering very important but finding optimal order is NP-hard as well 2

3 Query Vertex Ordering Query Graph: Data Graph: A u 1 A B B C C B C C C D u 2 E u 5 D D E E E F u 3 u 4 u 6 F F QuickSI: (u 1, u 2, u 3, u 4, u 5, u 6 ) 3

4 Query Vertex Ordering Query Graph: Data Graph: A u 1 A B B C C B C C C D u 2 E u 5 D D E E E F u 3 u 4 u 6 F F QuickSI: (u 1, u 2, u 3, u 4, u 5, u 6 ) 4

5 Core-Forest-Leaf Decomposition Core: dense subgraph of Query Graph A u 1 B C u 2 u 5 Forest: number of trees connected to B C Core via one vertex D u 2 u 5 u 3 Leafs: leafs of trees in the Forest structure D u 3 C u 5 F E u 4 u 6 5

6 Core-Forest-Leaf Decomposition A u 1 B u 2 C u 5 D u 3 E u 6 F u 4 matching order: Core Forest Leafs 6

7 Compact Path Index Idea Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B u 1 v 1 u 2 v 2 v 3 u 5 v 4 v 5 v

8 Compact Path Index - Construction criteria for data vertices in candidate set of a query vertex: 1. matching label 2. neighbor in candidate set of each already processed neighbor query vertex 8

9 Compact Path Index Construction Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 9

10 Compact Path Index Construction u 1 v 1 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 10

11 Compact Path Index Construction u 1 v 1 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 11

12 Compact Path Index Construction u 1 v 1 u 2 v 2 v 3 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 12

13 Compact Path Index Construction u 1 v 1 u 2 v 2 v 3 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 13

14 Compact Path Index Construction u 1 v 1 u 2 v 2 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 14

15 Compact Path Index Construction u 1 v 1 u 2 v 2 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 15

16 Compact Path Index Construction u 1 v 1 u 2 v 2 u 5 v 4 v 5 v 1004 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 16

17 Compact Path Index Construction u 1 v 1 u 2 v 2 u 5 v 4 v 5 v 1004 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 17

18 Compact Path Index Construction u 1 v 1 u 2 v 2 u 5 v 4 v 5 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 18

19 Compact Path Index - Application Query Graph: Index: A u 1 u 1 v 1 B C u 2 u 5 u 2 v 2 u 5 v 4 v 5 P 1 : 1 embedding P 2 : 2 embeddings A u 1 A u 1 B C u 2 u 5 19

20 Compact Path Index - Application matching sequence: start with P 1 = (u 1, u 2 ), add P 2 = (u 1, u 5 ) S = (u 1, u 2, u 5 ) iteratively map query vertices to candidates in the index, check data graph for non-tree edges P 1 : 1 embedding P 2 : 2 embeddings A u 1 A u 1 B C u 2 u 5 20

21 Improved Query Vertex Ordering Query Graph: Data Graph: A u 1 A B B C C B C C C D u 2 E u 5 D D E E E F u 3 u 4 u 6 F F CFL: (u 1, u 2, u 5, u 3, u 4, u 6 ) 21

22 Evaluation compared to QuickSI and TurboISO several datasets of varying size ( nodes) real Protein Interaction Networks synthetic graphs several generated queries of varying size ( nodes) 22

23 Evaluation faster by multiple orders of magnitude 23

24 References Bi, Fei, et al. "Efficient subgraph matching by postponing cartesian products." Proceedings of the 2016 International Conference on Management of Data. ACM,

25 Supervised Random Walks Predicting and Recommending Links in Social Networks Ronny Pachel Uni Potsdam Feb 07th 2017 Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

26 Contents Introduction 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

27 Link Prediction Introduction Network consisting of nodes and edges Between which two nodes will new edges (links) be created? [In a specific timeframe] Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

28 Challenges Introduction Motivation Author working at Facebook - motivation obvious Every new Link (Friendships, etc.) is increasing Network value Challenges of real Networks Extremely sparse Facebook has nearly 2 billion users, how many friends do you have? To predict that no new Links will be create is nearly 100% accurate! What network features have an influence on the creation of new links? Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

29 Present Work Introduction Method for link prediction and recommendation with supervised random walks Combines network structure and network properties PageRank-like random walk that visits given nodes (positive training examples) more often LEARN edge strength (transition probabilities) Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

30 Contents Supervised Random Walks 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

31 Random Walks Supervised Random Walks Random Walks! 0.4 B A C Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

32 Supervised Random Walks Transition Matrix Probability after one step: = Stationary Probability: /3 1/ /3 = 1/ /3 1/3 Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

33 Overview Supervised Random Walks Given is a graph G Starting from a single node s with already existing Links and for which new links should be predicted Also required is a set of positive and negative training Nodes General approach - can also be used for link prediction, recommendation, link anomaly detection, missing link prediction, etc. Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

34 Approaches Supervised Random Walks We have two general approaches for this problem: As Classification task: Each candidate node is either a created Link in the training data (positive example), or not (negative example) But: Has high class imbalance, which makes learning very hard! How to take the structure of the Network into account? And: The choice and extraction of useful features is very hard! There are countless ways to describe node proximity alone! As Ranking Task: Give Nodes with new created Links a higher Score For Example PageRank and Random Walks with Restarts (RWR) are used Stationary Distribution of a Random Walk gives a measure of closeness Takes the structure into account, but not the properties Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

35 This Approach Supervised Random Walks Combination of Both Approaches - Uses Network Structure and Node and Edge Properties Idea: Control a RWR via its transisiton properties to visit positive training nodes more often than the negative training nodes Trivial Approach by setting the transition properties manually would massively overfit Solution: Find transition properties depending on node and edge properties via a learning algorithm! Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

36 Supervised Random Walks Problem Formulation Given Directed Graph: G(V, E) Start node: s Set of candidate nodes: C = c i = D L Set of destination nodes (positive training examples): D = d i Set of no-link nodes (negative training examples): L = l i Feature Vector for each Edge (u, v): ψ uv Includes the Properties of both, the Nodes and the Edge! Find Edge Strength: a uv Used as transition probability Use Learned Function in Training phase f w (ψ uv ) w: vector of weights Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

37 Supervised Random Walks Predicting Process We take a source node for which we want to predict future links For each Edge (u, v) in G we compute the strength a uv = f w (ψ u v) Then a RWR is run and stationary distribution assigns every node u a probability value p u Nodes are ordered by p u values - highest values are predicted as destinations for future links Weights must be learned! Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

38 Supervised Random Walks Optimization Problem min w F (w) = w 2 + λ d D,l L h(p l p d ) Minimization Problem! p d should be higher than p l h(): loss function - measures how bad the current prediction is in form of a penalty w 2 : second norm of the weight vector - smaller values are wanted! λ: trade-off between the loss and the weights vector A high λ means that the loss is more important than small weights, used to prevent overfitting to the data But how do we come from the input (w) to the measurable output (p)? Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

39 Supervised Random Walks Optimization Problem Set a start value of w For each edge (u,v) calculate the edge transition a uv = f (ψ uv ) Build the random walk transition probability matrix Q Incorporate the restart probabability α into Q Find the vector p of the node probabilities (also known as Personalized PageRank) via a eigenvalue problem (or stationary distribution) This process can be written as a single formula The derivation of this formula is used for a iterative calculation to converge to the p vector (gradient descent method) If I change the weight values like this, is the final result (F(w)) better (smaller)? Note: This problem is not generally convex - no global minimum is guaranteed! Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

40 Contents Experimental Setup 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

41 Network Experimental Setup 4 physics co-authorship networks, complete Facebook network of Iceland Iceland has a very high Facebook penetration Relatively few Edges link to users from other countries Timeframe from Nov 1st 2009 (t 1 ) to Jun 13th 2010 (t 2 ) More than users, about 55% of Icelands population On average 168 friends at t 1 and 26 new friends until t 2 Focus on predicting Links to Nodes that are 2 hops away Empirically most new Links close a triangle Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

42 Preprocessing Experimental Setup Only active Nodes with more than 20 Links Candidate Nodes that have less than 4 common friends have been pruned Empirically shown that it is very unlikely that they would become friends Nodes with too many links have been pruned Even after just 2 hops some nodes can have several millions neighbours These measure speed up the computation but do not influence the prediction performance much Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

43 Learner Setup Experimental Setup From these users, 200 have been randomly selected, 100 as training set to learn the algorithm, and 100 as test set to measure the performance These Properties have been used to generate 7 Features: Edge Age Edge Initiator Communication and Observation Features Number of Common Friends Two performance metrics have been used: AUC - Area under ROC curve Prec@20 - Precision at the Top 20 Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

44 Contents Experiments on Facebook Data 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

45 Algorithm Experiments on Facebook Data We need to choose 4 aspects of the algorithm: Loss Function Edge Strength Function Choice of RWR Restart Parameter α Choice of Regularization Parameter λ Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

46 Experiments on Facebook Data Loss Function Must be continuous and differentiable to optimize over them candidates: Squared loss with margin b: h(x) = max{x + b, 0} 2 Huber Loss with margin b and window z b: 0 if x b, h(x) = (x + b) 2 /(2z) if b < x z b, (x + b) z/2 if x > z b Wilcoxon-Mann-Whitney (WMW) loss with width b: 1 h(x) = 1+exp( x/b) Squared loss and Huber loss can be computed easier WMW has a slightly better performance But: Only WMW increases the AUC and Prec@20! Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

47 Experiments on Facebook Data Strength Function f w (ψ u v) must be non-negative and differentiable Two candidates, both are basing on the inner product of the weight vector and the feature vector candidates: Exponential edge strength: a uv = exp(ψ uv w) Logistic edge strength: a uv = (1 + exp( ψ w)) 1 Experiments show no significant difference in performance But the exponential function has a potential problem with underflowing or overflowing double precision numbers Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

48 Choice of α Experiments on Facebook Data α controls how far a walk wanders before it s reset α = 0 : PageRank of a node is it s degree α = 1 : After one hop the walk is reset Big values are short and local walks, low values are long Have a influence in unweighted graphs, but with edge strengths assigned the influence diminishes Empirically 0.3 < α < 0.7 has same performance Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

49 Choice of λ Experiments on Facebook Data Overfitting is not an issue And number of features is small We chose a λ of 1 Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

50 Results Experiments on Facebook Data SRW with all weights w = 0 as starting point Compared to two unsupervised methods (RWR, node degree) and two supervised machine learning methods (decision trees and logistic regression) Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

51 Contents Conclusion 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

52 Conclusions Conclusion Supervised Random Walks - a new learning algorithm for link prediction and recommendation Combines two different approaches that see the problem either as classification task or as ranking task Demonstrated on real Facebook data Predictions show large improvements over RWR and easier than classical machine learning approaches SRW can be adopted to many other problems Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28

53 Social Influence Computation and Maximization in Signed Networks with Competing Cascades David Schumann Hasso Plattner Institute Graph Mining Tuesday 7 th February, / 12

54 What is this paper about? Domain: Network of susceptible nodes. Trust and Distrust Relationships Diffusion Model: Independent Cascade Problem: Starting seed-set of size m with the highest expectation of nodes colored red Signed Network Influence Maximization (SiNiMax) Problem NP-hard 2 / 12

55 What is this paper about? 3 / 12

56 Signed Networks with competing Cascades Weighted, directed and signed graph G = (V, E, W ) Weight Matrix W : sign represents trust or distrust relationships. Absolute value denotes strength of influence / 12

57 Signed Networks with competing Cascades Two influnce diffusion processes spreading in discrete time steps Using the standard Independent Cascade Model (ICM) Infected node v gets one chance to infect neighbour w with probability p v,w On success w becomes infected at step t + 1. Infected nodes never return to susceptible state / 12

58 Signed Networks with competing Cascades Two influnce diffusion processes spreading in discrete time steps Using the standard Independent Cascade Model (ICM) Infected node v gets one chance to infect neighbour w with probability p v,w On success w becomes infected at step t + 1. Infected nodes never return to susceptible state / 12

59 Signed Networks with competing Cascades Two influnce diffusion processes spreading in discrete time steps Using the standard Independent Cascade Model (ICM) Infected node v gets one chance to infect neighbour w with probability p v,w On success w becomes infected at step t + 1. Infected nodes never return to susceptible state / 12

60 Signed Networks with competing Cascades Two influnce diffusion processes spreading in discrete time steps Using the standard Independent Cascade Model (ICM) Infected node v gets one chance to infect neighbour w with probability p v,w On success w becomes infected at step t + 1. Infected nodes never return to susceptible state / 12

61 Help of the Unified Model Reduces computational cost for commonly used diffusion models. Using: Individual and Collective Influence. For ICM: p (v,u) (t) = 0 r u,t = 1 P (no infected neighbor succeeds) r u,t = 1 (1 p(b v,t 1 ))... v N i (u) 6 / 12

62 Unified Model of Competing Cascades r + u,t r u,t B + u,t B u,t = Collective influence on u to be red. = Collective influence on u to be blue. = Probability of u being red at or before time t. = Probability of u being blue at or before time t. B u,t + (k, +) = Probability of node u being colored red at time t if node k is infected with red at time t 1. 7 / 12

63 Online Seed-set Selection using the Unified Model for Signed Networks: OSSU M ± Algorithm 1 pseudocode for OSSUM± Input: G, m Output: S 1: S 2: for t = 1 to m do 3: (k, c) = arg max (k,c) u (B+ u,t (k, c) B+ u,t ) 4: S S (k, c) 5: end for Runtime Complexity: O(m( V + E )) 8 / 12

64 Experiment Results Comparison of OSSU M ± to three other heuristics: Positive Degree: Nodes with highest positive degree are selected Positive Degree Discount: Variation of Positive Degree Effective Degree: Nodes with highest effective degree are selected Data Set: Epinions Slashdot # nodes 131k 82k # edges 840k 550k # positive edges 710k 425k # negative edges 130k 125k 9 / 12

65 Experiment Results 10 / 12

66 Experiment Results 11 / 12

67 Review Positive: Introduced novel influence maximization problem SiNiMax OSSU M ± outperforms state-of-the-art heuristics for influence maximization Application to real world data Negative: Inaccuracies in description of method (missing edge cases) Unproven claim of extensaibility of this approach Real world data had to be modified to show advantage of OSSUM± 12 / 12

68 Top-k Link Recommendation in Social Graphs Sören Tietböhl

69

70

71

72

73 X = UV

74 X = UV

75

76 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6)

77 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases

78 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List:

79 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List: 1. Define new loss function that penalizes mistakes at the top more than at the bottom 3 4 6

80 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List: 1. Define new loss function that penalizes mistakes at the top more than at the bottom 2. Derive an algorithm that minimizes that function 3 4 6

81 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List: 1. Define new loss function that penalizes mistakes at the top more than at the bottom 2. Derive an algorithm that minimizes that function 3. Evaluate the algorithm using various datasets and metrics 3 4 6

82 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List: 1. Define new loss function that penalizes mistakes at the top more than at the bottom 2. Derive an algorithm that minimizes that function 3. Evaluate the algorithm using various datasets and metrics 4.??? 5. Profit

83

84

85 = Training Adjacency Matrix = Ranking Score

86 = Training Adjacency Matrix = Ranking Score (2,4),(4,5),(3,4),(1,4),(4,6)

87 = Training Adjacency Matrix = Ranking Score (2,4),(4,5),(3,4),(1,4),(4,6)

88 = Training Adjacency Matrix = Ranking Score (2,4),(4,5),(3,4),(1,4),(4,6)

89 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6)

90 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6)

91 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Prediction 1:

92 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Prediction 1: Prediction 2:

93

94

95 Sigmoid function:

96 Sigmoid function: Now its derivable!

97 Sigmoid function: Now its derivable! yay

98 Sigmoid function: Now its derivable!

99

100

101

102

103

104 (2,4),(3,4),(4,5),(1,4),(4,6)

105 (2,4),(3,4),(4,5),(1,4),(4,6) (1,4),(2,4),(3,4),(4,5),(4,6)

106

107

108

109

110 Sources Dongjin Song, David A. Meyer, Dacheng Tao, "Top-k Link Recommendation in Social Networks", 2015 IEEE International Conference on Data Mining (ICDM), vol. 00, no., pp , 2015, doi: /icdm Graph Mining Course Slides quicklatex.com

111 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel Graph Mining WT 2016/17 1 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

112 Structural Graph Clustering Cluster v 1 v 9 Hub v 2 v 4 v 5 v 6 v 8 Outlier v 3 v 7 v 10 2 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

113 Agenda Principles of SCAN Improvements of pscan Experimental Results 3 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

114 Principles of SCAN Graph G = (V, E) Structural Neighborhood: N u = {v V (u, v) E} {u} d u = N u Vertex u v 1 v 9 Neighborhood N u v 2 v 4 v 5 v 6 v 8 v 3 v 7 v 10 4 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

115 Principles of SCAN Structural Similarity: σ(u, v) = N u N v du d v Neighborhood N u v 1 v 9 Neighborhood N v v 2 v 4 v 5 v 6 v 8 N u N v v 3 v 7 σ(v 6, v 5 ) = v 10 5 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

116 Principles of SCAN Similarity Threshold and Core Vertices: Nu ɛ = {v N u σ(u, v) ɛ} V C = {v V Nv ɛ µ} Core Vertex ɛ-edge ɛ = 0.7 v v v v v v v 9 v v 8 µ = 4 v 10 6 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

117 Principles of SCAN Structure Reachability: u is structure reachable by v if there is a Path of ɛ-edges and intermediate core vertices Cluster C V such that all Vertices v C are mutually structure reachable and no Vertex x V \ C is structure reachable by any Vertex in C Core Vertex v 1 v 9 ɛ-edge v 2 v 4 v 5 v 6 v 8 ɛ = 0.7 v 3 v 7 µ = 4 v 10 7 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

118 Principles of SCAN Algorithm: SCAN requires structural similarity calculation along all edges in G equivalent to enumeration of all triangles in G complexity of O( E 1.5 ) which is worst case optimum 8 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

119 Agenda Principles of SCAN Improvements of pscan Experimental Results 9 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

120 Improvements of pscan Paradigm: 1 cluster core vertices 2 expand clusters by core vertices structural neighborhood Effective Degree and Similar Degree: vertex u V with incomplete similarity measures in its neighborhood effective degree: d e u = d u x where x is the number of neighbors v V with σ(u, v) < ɛ known similar degree: d s u number of neighbors v V with σ(u, v) ɛ known d e u and d s u provide an upper and lower bound for N ɛ u calculate structural neighborhoods for vertices in non-increasing d e -order update d e and d s for every calculated σ; omit neighborhood calculation if d e or d s are sufficient to identify core vertex 10 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

121 Improvements of pscan u ds u d e u v v v v v v v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

122 Improvements of pscan u ds u d e u v v v v 4 5 v 5 3 v 6 4 v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

123 Improvements of pscan u ds u d e u v v v v 4 4 v 5 2 v 6 4 v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

124 Improvements of pscan u ds u d e u v v v v 4 4 v 5 2 v v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

125 Improvements of pscan u ds u d e u v v v v 4 4 v 5 2 v v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

126 Improvements of pscan u ds u d e u 1 v v ɛ = 0.7 v v v 4 4 v 5 2 v v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

127 Improvements of pscan u ds u d e u 1 v v ɛ = 0.7 v v v 4 4 v 5 2 v v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

128 Improvements of pscan u ds u d e u 1 v v ɛ = 0.7 v v v 4 4 v 5 2 v v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

129 Agenda Principles of SCAN Improvements of pscan Experimental Results 19 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

130 Experimental Results Algorithms: SCAN SCAN++ (approximative version of SCAN omitting some similarity calculations) pscan (5 variants with different degrees of optimization) Data: 13 real world networks (13k to 1.2G edges) 11 generated networks with varying characteristics 20 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

131 Experimental Results 1 order of magnitude faster than either SCAN or SCAN++ 21 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17

132 Clustering Large Probabilistic Graphs Marianne Thieffry

133 Agenda 1. Definition of pclusteredit 2. Algorithms for pclusteredit 3. Generalizations of pclusteredit

134 Definition of pclusteredit

135 Probabilistic Graphs

136 Probabilistic Graphs

137 Probabilistic Graphs

138 Cluster graphs

139 Edit distance # edges added + # edges removed

140 Edit distance # edges added + # edges removed [between all deterministic instances of a given probabilistic graph and a given deterministic graph ]

141 pclusteredit [while minimizing the edit distance]

142 pclusteredit

143 CorrelationClustering

144 CorrelationClustering

145 CorrelationClustering

146 Algorithms for pclusteredit

147 pkwikcluster

148 pkwikcluster

149 pkwikcluster

150 pkwikcluster

151 pkwikcluster 0.8

152 pkwikcluster O(m)

153 Furthest

154 Furthest

155 Furthest

156 Furthest

157 Furthest O(nk²)

158 Agglomerative Singleton clusters Merge clusters with highest average edge probability, until highest average edge probability is at most 0.5 Complexity O(kn log n)

159 pkwikcluster

160 pkwikcluster

161 pkwikcluster

162 pkwikcluster

163 pkwikcluster O(kn log n)

164 Significance

165 Evaluation - CORE dataset protein-protein interaction network 2708 nodes 7123 edges

166 Evaluation - CORE dataset

167 Evaluation - CORE dataset

168 Evaluation - CORE dataset

169 Evaluation - CORE dataset

170 Generalizations of pclusteredit

171 Noisy cluster graphs

172 p2clusteredit & minα Find α-cluster with minimal edit distance Find α that produces minimum α-cluster with minimal edit distance

173 The shortest Path is not always a straight line From Vasiliki Kalavri, Tiago Simas & Dionysios Logothetis Philippe ZOU Graph Mining - WS 2016/2017 Davide Mottin & Konstantina Lazaridou 1

174 Agenda 1. Overview 2. Implementation 3. Evaluation 4. Conclusion 2

175 Overview 3

176 Motivations Leverage the Backbone metric The Blackbone metric is the minimum subgraph that preserves the shortest paths of a weighted graph Improve graph analysis tools and computing performance Large graph reduce running time and storage 4

177 Backbone Metric For instance, in the example they give : 5

178 Concept of semi-metricity An edge is semi-metric if you can find an indirect shorter path between the edges. Let consider the previous graph 6

179 Concept of semi-metricity An edge is semi-metric if you can find an indirect shorter path between the edges. Let consider the previous graph = 6 < 10 AD is semi-metric (same for the node CE) 7

180 Concept of semi-metricity Semi-metricity Degree : number of edges to add to reach the shortest path compared to the original (straight) one AD will be 2nd order (and CE will be 1st order) 8

181 Concept of semi-metricity Performed an analysis on data sets 9

182 Implementation 10

183 Backbone algorithm 3 phases Find 1st order semi-metric edges Only look at triangles 11

184 Backbone algorithm 3 phases Find 1st order semi-metric edges Only look at triangles 12

185 Backbone algorithm 3 phases Focus on 2nd order semi-metric edges 13

186 Backbone algorithm 3 phases Focus on 2nd order semi-metric edges Lowest weight edge of every node is metric? 14

187 Backbone algorithm 3 phases Breadth first search? 15

188 Backbone algorithm 3 phases Breadth first search? 16

189 Backbone algorithm 3 phases Breadth first search? 17

190 Backbone algorithm 3 phases Breadth first search? Target reached but with a shorter path -> Semi-metric edge 18

191 Backbone algorithm 3 phases End of algorithm 19

192 Evaluation 20

193 Execution time for phase 1 Execution time for finding triangles + removal 1st order semi-metric 21

194 Execution time for phase 1 Increase linearly. 15 Minutes for 4 Billions edges. 22

195 Storage Size reduction of the network after removal 23

196 Conclusion 24

197 Conclusion Provide a new concept in graphs : semi-metricity Algorithm to lower the number of edges in order to improve computation and reduce time show that direct nodes are not necessarily the shortest ones Challenge the large graphes issue 25

Link Prediction. Eman Badr Mohammed Saquib Akmal Khan

Link Prediction. Eman Badr Mohammed Saquib Akmal Khan Link Prediction Eman Badr Mohammed Saquib Akmal Khan 11-06-2013 Link Prediction Which pair of nodes should be connected? Applications Facebook friend suggestion Recommendation systems Monitoring and controlling

More information

Overlapping Community Detection at Scale: A Nonnegative Matrix Factorization Approach

Overlapping Community Detection at Scale: A Nonnegative Matrix Factorization Approach Overlapping Community Detection at Scale: A Nonnegative Matrix Factorization Approach Author: Jaewon Yang, Jure Leskovec 1 1 Venue: WSDM 2013 Presenter: Yupeng Gu 1 Stanford University 1 Background Community

More information

Logistic Regression Logistic

Logistic Regression Logistic Case Study 1: Estimating Click Probabilities L2 Regularization for Logistic Regression Machine Learning/Statistics for Big Data CSE599C1/STAT592, University of Washington Carlos Guestrin January 10 th,

More information

More on NP and Reductions

More on NP and Reductions Indian Institute of Information Technology Design and Manufacturing, Kancheepuram Chennai 600 127, India An Autonomous Institute under MHRD, Govt of India http://www.iiitdm.ac.in COM 501 Advanced Data

More information

Computing Competing Cascades on Signed Networks

Computing Competing Cascades on Signed Networks Noname manuscript No. (will be inserted by the editor Computing Competing Cascades on Signed Networks Ajitesh Srivastava Charalampos Chelmis Viktor K. Prasanna Received: date / Accepted: date Abstract

More information

Determining the Diameter of Small World Networks

Determining the Diameter of Small World Networks Determining the Diameter of Small World Networks Frank W. Takes & Walter A. Kosters Leiden University, The Netherlands CIKM 2011 October 2, 2011 Glasgow, UK NWO COMPASS project (grant #12.0.92) 1 / 30

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

Tufts COMP 135: Introduction to Machine Learning

Tufts COMP 135: Introduction to Machine Learning Tufts COMP 135: Introduction to Machine Learning https://www.cs.tufts.edu/comp/135/2019s/ Logistic Regression Many slides attributable to: Prof. Mike Hughes Erik Sudderth (UCI) Finale Doshi-Velez (Harvard)

More information

Web Structure Mining Nodes, Links and Influence

Web Structure Mining Nodes, Links and Influence Web Structure Mining Nodes, Links and Influence 1 Outline 1. Importance of nodes 1. Centrality 2. Prestige 3. Page Rank 4. Hubs and Authority 5. Metrics comparison 2. Link analysis 3. Influence model 1.

More information

CSCI-567: Machine Learning (Spring 2019)

CSCI-567: Machine Learning (Spring 2019) CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Mar. 19, 2019 March 19, 2019 1 / 43 Administration March 19, 2019 2 / 43 Administration TA3 is due this week March

More information

Predictive analysis on Multivariate, Time Series datasets using Shapelets

Predictive analysis on Multivariate, Time Series datasets using Shapelets 1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,

More information

Introduction to Logistic Regression

Introduction to Logistic Regression Introduction to Logistic Regression Guy Lebanon Binary Classification Binary classification is the most basic task in machine learning, and yet the most frequent. Binary classifiers often serve as the

More information

Greedy Maximization Framework for Graph-based Influence Functions

Greedy Maximization Framework for Graph-based Influence Functions Greedy Maximization Framework for Graph-based Influence Functions Edith Cohen Google Research Tel Aviv University HotWeb '16 1 Large Graphs Model relations/interactions (edges) between entities (nodes)

More information

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan Clustering CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Supervised vs Unsupervised Learning Supervised learning Given x ", y " "%& ', learn a function f: X Y Categorical output classification

More information

CS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine

CS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine CS 277: Data Mining Mining Web Link Structure Class Presentations In-class, Tuesday and Thursday next week 2-person teams: 6 minutes, up to 6 slides, 3 minutes/slides each person 1-person teams 4 minutes,

More information

Protein Complex Identification by Supervised Graph Clustering

Protein Complex Identification by Supervised Graph Clustering Protein Complex Identification by Supervised Graph Clustering Yanjun Qi 1, Fernanda Balem 2, Christos Faloutsos 1, Judith Klein- Seetharaman 1,2, Ziv Bar-Joseph 1 1 School of Computer Science, Carnegie

More information

Ad Placement Strategies

Ad Placement Strategies Case Study : Estimating Click Probabilities Intro Logistic Regression Gradient Descent + SGD AdaGrad Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox January 7 th, 04 Ad

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Overlapping Communities

Overlapping Communities Overlapping Communities Davide Mottin HassoPlattner Institute Graph Mining course Winter Semester 2017 Acknowledgements Most of this lecture is taken from: http://web.stanford.edu/class/cs224w/slides GRAPH

More information

Case Study 1: Estimating Click Probabilities. Kakade Announcements: Project Proposals: due this Friday!

Case Study 1: Estimating Click Probabilities. Kakade Announcements: Project Proposals: due this Friday! Case Study 1: Estimating Click Probabilities Intro Logistic Regression Gradient Descent + SGD Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade April 4, 017 1 Announcements:

More information

Summary and discussion of: Dropout Training as Adaptive Regularization

Summary and discussion of: Dropout Training as Adaptive Regularization Summary and discussion of: Dropout Training as Adaptive Regularization Statistics Journal Club, 36-825 Kirstin Early and Calvin Murdock November 21, 2014 1 Introduction Multi-layered (i.e. deep) artificial

More information

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning Mark Schmidt University of British Columbia, May 2016 www.cs.ubc.ca/~schmidtm/svan16 Some images from this lecture are

More information

Machine Learning for NLP

Machine Learning for NLP Machine Learning for NLP Linear Models Joakim Nivre Uppsala University Department of Linguistics and Philology Slides adapted from Ryan McDonald, Google Research Machine Learning for NLP 1(26) Outline

More information

Communities Via Laplacian Matrices. Degree, Adjacency, and Laplacian Matrices Eigenvectors of Laplacian Matrices

Communities Via Laplacian Matrices. Degree, Adjacency, and Laplacian Matrices Eigenvectors of Laplacian Matrices Communities Via Laplacian Matrices Degree, Adjacency, and Laplacian Matrices Eigenvectors of Laplacian Matrices The Laplacian Approach As with betweenness approach, we want to divide a social graph into

More information

CPSC 340: Machine Learning and Data Mining

CPSC 340: Machine Learning and Data Mining CPSC 340: Machine Learning and Data Mining Linear Classifiers: predictions Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due Friday of next

More information

Training the linear classifier

Training the linear classifier 215, Training the linear classifier A natural way to train the classifier is to minimize the number of classification errors on the training data, i.e. choosing w so that the training error is minimized.

More information

Gradient Boosting (Continued)

Gradient Boosting (Continued) Gradient Boosting (Continued) David Rosenberg New York University April 4, 2016 David Rosenberg (New York University) DS-GA 1003 April 4, 2016 1 / 31 Boosting Fits an Additive Model Boosting Fits an Additive

More information

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS DATA MINING LECTURE 3 Link Analysis Ranking PageRank -- Random walks HITS How to organize the web First try: Manually curated Web Directories How to organize the web Second try: Web Search Information

More information

Data Mining and Analysis: Fundamental Concepts and Algorithms

Data Mining and Analysis: Fundamental Concepts and Algorithms Data Mining and Analysis: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table

More information

Linear classifiers: Overfitting and regularization

Linear classifiers: Overfitting and regularization Linear classifiers: Overfitting and regularization Emily Fox University of Washington January 25, 2017 Logistic regression recap 1 . Thus far, we focused on decision boundaries Score(x i ) = w 0 h 0 (x

More information

18.9 SUPPORT VECTOR MACHINES

18.9 SUPPORT VECTOR MACHINES 744 Chapter 8. Learning from Examples is the fact that each regression problem will be easier to solve, because it involves only the examples with nonzero weight the examples whose kernels overlap the

More information

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring /

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring / Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring 2015 http://ce.sharif.edu/courses/93-94/2/ce717-1 / Agenda Combining Classifiers Empirical view Theoretical

More information

Midterm Exam, Spring 2005

Midterm Exam, Spring 2005 10-701 Midterm Exam, Spring 2005 1. Write your name and your email address below. Name: Email address: 2. There should be 15 numbered pages in this exam (including this cover sheet). 3. Write your name

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table

More information

A New Space for Comparing Graphs

A New Space for Comparing Graphs A New Space for Comparing Graphs Anshumali Shrivastava and Ping Li Cornell University and Rutgers University August 18th 2014 Anshumali Shrivastava and Ping Li ASONAM 2014 August 18th 2014 1 / 38 Main

More information

Overfitting, Bias / Variance Analysis

Overfitting, Bias / Variance Analysis Overfitting, Bias / Variance Analysis Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 8, 207 / 40 Outline Administration 2 Review of last lecture 3 Basic

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Capacity Releasing Diffusion for Speed and Locality

Capacity Releasing Diffusion for Speed and Locality 000 00 002 003 004 005 006 007 008 009 00 0 02 03 04 05 06 07 08 09 020 02 022 023 024 025 026 027 028 029 030 03 032 033 034 035 036 037 038 039 040 04 042 043 044 045 046 047 048 049 050 05 052 053 054

More information

The exam is closed book, closed notes except your one-page cheat sheet.

The exam is closed book, closed notes except your one-page cheat sheet. CS 189 Fall 2015 Introduction to Machine Learning Final Please do not turn over the page before you are instructed to do so. You have 2 hours and 50 minutes. Please write your initials on the top-right

More information

Unsupervised Learning

Unsupervised Learning 2018 EE448, Big Data Mining, Lecture 7 Unsupervised Learning Weinan Zhang Shanghai Jiao Tong University http://wnzhang.net http://wnzhang.net/teaching/ee448/index.html ML Problem Setting First build and

More information

Max Margin-Classifier

Max Margin-Classifier Max Margin-Classifier Oliver Schulte - CMPT 726 Bishop PRML Ch. 7 Outline Maximum Margin Criterion Math Maximizing the Margin Non-Separable Data Kernels and Non-linear Mappings Where does the maximization

More information

Kristina Lerman USC Information Sciences Institute

Kristina Lerman USC Information Sciences Institute Rethinking Network Structure Kristina Lerman USC Information Sciences Institute Università della Svizzera Italiana, December 16, 2011 Measuring network structure Central nodes Community structure Strength

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu March 16, 2016 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision

More information

SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels

SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels Karl Stratos June 21, 2018 1 / 33 Tangent: Some Loose Ends in Logistic Regression Polynomial feature expansion in logistic

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 622 - Section 2 - Spring 27 Pre-final Review Jan-Willem van de Meent Feedback Feedback https://goo.gl/er7eo8 (also posted on Piazza) Also, please fill out your TRACE evaluations!

More information

Learning theory. Ensemble methods. Boosting. Boosting: history

Learning theory. Ensemble methods. Boosting. Boosting: history Learning theory Probability distribution P over X {0, 1}; let (X, Y ) P. We get S := {(x i, y i )} n i=1, an iid sample from P. Ensemble methods Goal: Fix ɛ, δ (0, 1). With probability at least 1 δ (over

More information

Online Advertising is Big Business

Online Advertising is Big Business Online Advertising Online Advertising is Big Business Multiple billion dollar industry $43B in 2013 in USA, 17% increase over 2012 [PWC, Internet Advertising Bureau, April 2013] Higher revenue in USA

More information

Regression with Numerical Optimization. Logistic

Regression with Numerical Optimization. Logistic CSG220 Machine Learning Fall 2008 Regression with Numerical Optimization. Logistic regression Regression with Numerical Optimization. Logistic regression based on a document by Andrew Ng October 3, 204

More information

Point-of-Interest Recommendations: Learning Potential Check-ins from Friends

Point-of-Interest Recommendations: Learning Potential Check-ins from Friends Point-of-Interest Recommendations: Learning Potential Check-ins from Friends Huayu Li, Yong Ge +, Richang Hong, Hengshu Zhu University of North Carolina at Charlotte + University of Arizona Hefei University

More information

Research Overview. Kristjan Greenewald. February 2, University of Michigan - Ann Arbor

Research Overview. Kristjan Greenewald. February 2, University of Michigan - Ann Arbor Research Overview Kristjan Greenewald University of Michigan - Ann Arbor February 2, 2016 2/17 Background and Motivation Want efficient statistical modeling of high-dimensional spatio-temporal data with

More information

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation. CS 189 Spring 2015 Introduction to Machine Learning Midterm You have 80 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. No calculators or electronic items.

More information

Classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012 Classification CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Topics Discriminant functions Logistic regression Perceptron Generative models Generative vs. discriminative

More information

Support Vector Machines

Support Vector Machines Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized

More information

ECEN 689 Special Topics in Data Science for Communications Networks

ECEN 689 Special Topics in Data Science for Communications Networks ECEN 689 Special Topics in Data Science for Communications Networks Nick Duffield Department of Electrical & Computer Engineering Texas A&M University Lecture 8 Random Walks, Matrices and PageRank Graphs

More information

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology

More information

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 Exam policy: This exam allows two one-page, two-sided cheat sheets; No other materials. Time: 2 hours. Be sure to write your name and

More information

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted

More information

Linear and Logistic Regression. Dr. Xiaowei Huang

Linear and Logistic Regression. Dr. Xiaowei Huang Linear and Logistic Regression Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Two Classical Machine Learning Algorithms Decision tree learning K-nearest neighbor Model Evaluation Metrics

More information

Collaborative Filtering. Radek Pelánek

Collaborative Filtering. Radek Pelánek Collaborative Filtering Radek Pelánek 2017 Notes on Lecture the most technical lecture of the course includes some scary looking math, but typically with intuitive interpretation use of standard machine

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu 10/24/2012 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu

More information

Decision Trees: Overfitting

Decision Trees: Overfitting Decision Trees: Overfitting Emily Fox University of Washington January 30, 2017 Decision tree recap Loan status: Root 22 18 poor 4 14 Credit? Income? excellent 9 0 3 years 0 4 Fair 9 4 Term? 5 years 9

More information

Spectral Methods for Subgraph Detection

Spectral Methods for Subgraph Detection Spectral Methods for Subgraph Detection Nadya T. Bliss & Benjamin A. Miller Embedded and High Performance Computing Patrick J. Wolfe Statistics and Information Laboratory Harvard University 12 July 2010

More information

Submodular Functions Properties Algorithms Machine Learning

Submodular Functions Properties Algorithms Machine Learning Submodular Functions Properties Algorithms Machine Learning Rémi Gilleron Inria Lille - Nord Europe & LIFL & Univ Lille Jan. 12 revised Aug. 14 Rémi Gilleron (Mostrare) Submodular Functions Jan. 12 revised

More information

PATTERN CLASSIFICATION

PATTERN CLASSIFICATION PATTERN CLASSIFICATION Second Edition Richard O. Duda Peter E. Hart David G. Stork A Wiley-lnterscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane Singapore Toronto CONTENTS

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Learning Tetris. 1 Tetris. February 3, 2009

Learning Tetris. 1 Tetris. February 3, 2009 Learning Tetris Matt Zucker Andrew Maas February 3, 2009 1 Tetris The Tetris game has been used as a benchmark for Machine Learning tasks because its large state space (over 2 200 cell configurations are

More information

Performance Evaluation. Analyzing competitive influence maximization problems with partial information: An approximation algorithmic framework

Performance Evaluation. Analyzing competitive influence maximization problems with partial information: An approximation algorithmic framework Performance Evaluation ( ) Contents lists available at ScienceDirect Performance Evaluation journal homepage: www.elsevier.com/locate/peva Analyzing competitive influence maximization problems with partial

More information

Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups

Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups Contemporary Mathematics Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups Robert M. Haralick, Alex D. Miasnikov, and Alexei G. Myasnikov Abstract. We review some basic methodologies

More information

Using R for Iterative and Incremental Processing

Using R for Iterative and Incremental Processing Using R for Iterative and Incremental Processing Shivaram Venkataraman, Indrajit Roy, Alvin AuYoung, Robert Schreiber UC Berkeley and HP Labs UC BERKELEY Big Data, Complex Algorithms PageRank (Dominant

More information

Final Overview. Introduction to ML. Marek Petrik 4/25/2017

Final Overview. Introduction to ML. Marek Petrik 4/25/2017 Final Overview Introduction to ML Marek Petrik 4/25/2017 This Course: Introduction to Machine Learning Build a foundation for practice and research in ML Basic machine learning concepts: max likelihood,

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu November 16, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision

More information

Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning

Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning Nicolas Thome Prenom.Nom@cnam.fr http://cedric.cnam.fr/vertigo/cours/ml2/ Département Informatique Conservatoire

More information

Anomaly Detection. Davide Mottin, Konstantina Lazaridou. HassoPlattner Institute. Graph Mining course Winter Semester 2016

Anomaly Detection. Davide Mottin, Konstantina Lazaridou. HassoPlattner Institute. Graph Mining course Winter Semester 2016 Anomaly Detection Davide Mottin, Konstantina Lazaridou HassoPlattner Institute Graph Mining course Winter Semester 2016 and Next week February 7, third round presentations Slides are due by February 6,

More information

Analytically tractable processes on networks

Analytically tractable processes on networks University of California San Diego CERTH, 25 May 2011 Outline Motivation 1 Motivation Networks Random walk and Consensus Epidemic models Spreading processes on networks 2 Networks Motivation Networks Random

More information

CPSC 340: Machine Learning and Data Mining. Gradient Descent Fall 2016

CPSC 340: Machine Learning and Data Mining. Gradient Descent Fall 2016 CPSC 340: Machine Learning and Data Mining Gradient Descent Fall 2016 Admin Assignment 1: Marks up this weekend on UBC Connect. Assignment 2: 3 late days to hand it in Monday. Assignment 3: Due Wednesday

More information

Statistical Learning. Dong Liu. Dept. EEIS, USTC

Statistical Learning. Dong Liu. Dept. EEIS, USTC Statistical Learning Dong Liu Dept. EEIS, USTC Chapter 6. Unsupervised and Semi-Supervised Learning 1. Unsupervised learning 2. k-means 3. Gaussian mixture model 4. Other approaches to clustering 5. Principle

More information

Anomaly Detection for the CERN Large Hadron Collider injection magnets

Anomaly Detection for the CERN Large Hadron Collider injection magnets Anomaly Detection for the CERN Large Hadron Collider injection magnets Armin Halilovic KU Leuven - Department of Computer Science In cooperation with CERN 2018-07-27 0 Outline 1 Context 2 Data 3 Preprocessing

More information

Logistic Regression & Neural Networks

Logistic Regression & Neural Networks Logistic Regression & Neural Networks CMSC 723 / LING 723 / INST 725 Marine Carpuat Slides credit: Graham Neubig, Jacob Eisenstein Logistic Regression Perceptron & Probabilities What if we want a probability

More information

Empirical Risk Minimization

Empirical Risk Minimization Empirical Risk Minimization Fabrice Rossi SAMM Université Paris 1 Panthéon Sorbonne 2018 Outline Introduction PAC learning ERM in practice 2 General setting Data X the input space and Y the output space

More information

Is the test error unbiased for these programs?

Is the test error unbiased for these programs? Is the test error unbiased for these programs? Xtrain avg N o Preprocessing by de meaning using whole TEST set 2017 Kevin Jamieson 1 Is the test error unbiased for this program? e Stott see non for f x

More information

Learning from Examples

Learning from Examples Learning from Examples Data fitting Decision trees Cross validation Computational learning theory Linear classifiers Neural networks Nonparametric methods: nearest neighbor Support vector machines Ensemble

More information

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet.

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet. CS 189 Spring 013 Introduction to Machine Learning Final You have 3 hours for the exam. The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet. Please

More information

DATA MINING AND MACHINE LEARNING

DATA MINING AND MACHINE LEARNING DATA MINING AND MACHINE LEARNING Lecture 5: Regularization and loss functions Lecturer: Simone Scardapane Academic Year 2016/2017 Table of contents Loss functions Loss functions for regression problems

More information

hsnim: Hyper Scalable Network Inference Machine for Scale-Free Protein-Protein Interaction Networks Inference

hsnim: Hyper Scalable Network Inference Machine for Scale-Free Protein-Protein Interaction Networks Inference CS 229 Project Report (TR# MSB2010) Submitted 12/10/2010 hsnim: Hyper Scalable Network Inference Machine for Scale-Free Protein-Protein Interaction Networks Inference Muhammad Shoaib Sehgal Computer Science

More information

Clustering. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms March 8, / 26

Clustering. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms March 8, / 26 Clustering Professor Ameet Talwalkar Professor Ameet Talwalkar CS26 Machine Learning Algorithms March 8, 217 1 / 26 Outline 1 Administration 2 Review of last lecture 3 Clustering Professor Ameet Talwalkar

More information

Variations of Logistic Regression with Stochastic Gradient Descent

Variations of Logistic Regression with Stochastic Gradient Descent Variations of Logistic Regression with Stochastic Gradient Descent Panqu Wang(pawang@ucsd.edu) Phuc Xuan Nguyen(pxn002@ucsd.edu) January 26, 2012 Abstract In this paper, we extend the traditional logistic

More information

Multiclass Classification-1

Multiclass Classification-1 CS 446 Machine Learning Fall 2016 Oct 27, 2016 Multiclass Classification Professor: Dan Roth Scribe: C. Cheng Overview Binary to multiclass Multiclass SVM Constraint classification 1 Introduction Multiclass

More information

Support vector machines Lecture 4

Support vector machines Lecture 4 Support vector machines Lecture 4 David Sontag New York University Slides adapted from Luke Zettlemoyer, Vibhav Gogate, and Carlos Guestrin Q: What does the Perceptron mistake bound tell us? Theorem: The

More information

Data Mining Part 5. Prediction

Data Mining Part 5. Prediction Data Mining Part 5. Prediction 5.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline How the Brain Works Artificial Neural Networks Simple Computing Elements Feed-Forward Networks Perceptrons (Single-layer,

More information

Using Local Spectral Methods in Theory and in Practice

Using Local Spectral Methods in Theory and in Practice Using Local Spectral Methods in Theory and in Practice Michael W. Mahoney ICSI and Dept of Statistics, UC Berkeley ( For more info, see: http: // www. stat. berkeley. edu/ ~ mmahoney or Google on Michael

More information

A Gradient-based Adaptive Learning Framework for Efficient Personal Recommendation

A Gradient-based Adaptive Learning Framework for Efficient Personal Recommendation A Gradient-based Adaptive Learning Framework for Efficient Personal Recommendation Yue Ning 1 Yue Shi 2 Liangjie Hong 2 Huzefa Rangwala 3 Naren Ramakrishnan 1 1 Virginia Tech 2 Yahoo Research. Yue Shi

More information

Nearest Neighbor. Machine Learning CSE546 Kevin Jamieson University of Washington. October 26, Kevin Jamieson 2

Nearest Neighbor. Machine Learning CSE546 Kevin Jamieson University of Washington. October 26, Kevin Jamieson 2 Nearest Neighbor Machine Learning CSE546 Kevin Jamieson University of Washington October 26, 2017 2017 Kevin Jamieson 2 Some data, Bayes Classifier Training data: True label: +1 True label: -1 Optimal

More information

6.207/14.15: Networks Lecture 7: Search on Networks: Navigation and Web Search

6.207/14.15: Networks Lecture 7: Search on Networks: Navigation and Web Search 6.207/14.15: Networks Lecture 7: Search on Networks: Navigation and Web Search Daron Acemoglu and Asu Ozdaglar MIT September 30, 2009 1 Networks: Lecture 7 Outline Navigation (or decentralized search)

More information

2018 EE448, Big Data Mining, Lecture 4. (Part I) Weinan Zhang Shanghai Jiao Tong University

2018 EE448, Big Data Mining, Lecture 4. (Part I) Weinan Zhang Shanghai Jiao Tong University 2018 EE448, Big Data Mining, Lecture 4 Supervised Learning (Part I) Weinan Zhang Shanghai Jiao Tong University http://wnzhang.net http://wnzhang.net/teaching/ee448/index.html Content of Supervised Learning

More information

Generative v. Discriminative classifiers Intuition

Generative v. Discriminative classifiers Intuition Logistic Regression (Continued) Generative v. Discriminative Decision rees Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University January 31 st, 2007 2005-2007 Carlos Guestrin 1 Generative

More information

Introduction to Machine Learning. Introduction to ML - TAU 2016/7 1

Introduction to Machine Learning. Introduction to ML - TAU 2016/7 1 Introduction to Machine Learning Introduction to ML - TAU 2016/7 1 Course Administration Lecturers: Amir Globerson (gamir@post.tau.ac.il) Yishay Mansour (Mansour@tau.ac.il) Teaching Assistance: Regev Schweiger

More information

A Randomized Approach for Crowdsourcing in the Presence of Multiple Views

A Randomized Approach for Crowdsourcing in the Presence of Multiple Views A Randomized Approach for Crowdsourcing in the Presence of Multiple Views Presenter: Yao Zhou joint work with: Jingrui He - 1 - Roadmap Motivation Proposed framework: M2VW Experimental results Conclusion

More information

MODULE -4 BAYEIAN LEARNING

MODULE -4 BAYEIAN LEARNING MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities

More information

Maximizing the Spread of Influence through a Social Network. David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03

Maximizing the Spread of Influence through a Social Network. David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03 Maximizing the Spread of Influence through a Social Network David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03 Influence and Social Networks Economics, sociology, political science, etc. all have studied

More information