Efficient Subgraph Matching by Postponing Cartesian Products. Matthias Barkowsky
|
|
- Josephine Riley
- 6 years ago
- Views:
Transcription
1 Efficient Subgraph Matching by Postponing Cartesian Products Matthias Barkowsky 5
2 Subgraph Isomorphism - Approaches search problem: find all embeddings of a query graph in a data graph NP-hard, but can often be solved efficiently in practice general approach: iteratively map query vertices to data vertices -> ordering very important but finding optimal order is NP-hard as well 2
3 Query Vertex Ordering Query Graph: Data Graph: A u 1 A B B C C B C C C D u 2 E u 5 D D E E E F u 3 u 4 u 6 F F QuickSI: (u 1, u 2, u 3, u 4, u 5, u 6 ) 3
4 Query Vertex Ordering Query Graph: Data Graph: A u 1 A B B C C B C C C D u 2 E u 5 D D E E E F u 3 u 4 u 6 F F QuickSI: (u 1, u 2, u 3, u 4, u 5, u 6 ) 4
5 Core-Forest-Leaf Decomposition Core: dense subgraph of Query Graph A u 1 B C u 2 u 5 Forest: number of trees connected to B C Core via one vertex D u 2 u 5 u 3 Leafs: leafs of trees in the Forest structure D u 3 C u 5 F E u 4 u 6 5
6 Core-Forest-Leaf Decomposition A u 1 B u 2 C u 5 D u 3 E u 6 F u 4 matching order: Core Forest Leafs 6
7 Compact Path Index Idea Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B u 1 v 1 u 2 v 2 v 3 u 5 v 4 v 5 v
8 Compact Path Index - Construction criteria for data vertices in candidate set of a query vertex: 1. matching label 2. neighbor in candidate set of each already processed neighbor query vertex 8
9 Compact Path Index Construction Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 9
10 Compact Path Index Construction u 1 v 1 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 10
11 Compact Path Index Construction u 1 v 1 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 11
12 Compact Path Index Construction u 1 v 1 u 2 v 2 v 3 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 12
13 Compact Path Index Construction u 1 v 1 u 2 v 2 v 3 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 13
14 Compact Path Index Construction u 1 v 1 u 2 v 2 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 14
15 Compact Path Index Construction u 1 v 1 u 2 v 2 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 15
16 Compact Path Index Construction u 1 v 1 u 2 v 2 u 5 v 4 v 5 v 1004 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 16
17 Compact Path Index Construction u 1 v 1 u 2 v 2 u 5 v 4 v 5 v 1004 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 17
18 Compact Path Index Construction u 1 v 1 u 2 v 2 u 5 v 4 v 5 Query Graph: Data Graph: B u 2 A u 1 C u 5 v 1 A v 3 C B C C C v 4 v 2 v 5 v 6 v 1004 B 18
19 Compact Path Index - Application Query Graph: Index: A u 1 u 1 v 1 B C u 2 u 5 u 2 v 2 u 5 v 4 v 5 P 1 : 1 embedding P 2 : 2 embeddings A u 1 A u 1 B C u 2 u 5 19
20 Compact Path Index - Application matching sequence: start with P 1 = (u 1, u 2 ), add P 2 = (u 1, u 5 ) S = (u 1, u 2, u 5 ) iteratively map query vertices to candidates in the index, check data graph for non-tree edges P 1 : 1 embedding P 2 : 2 embeddings A u 1 A u 1 B C u 2 u 5 20
21 Improved Query Vertex Ordering Query Graph: Data Graph: A u 1 A B B C C B C C C D u 2 E u 5 D D E E E F u 3 u 4 u 6 F F CFL: (u 1, u 2, u 5, u 3, u 4, u 6 ) 21
22 Evaluation compared to QuickSI and TurboISO several datasets of varying size ( nodes) real Protein Interaction Networks synthetic graphs several generated queries of varying size ( nodes) 22
23 Evaluation faster by multiple orders of magnitude 23
24 References Bi, Fei, et al. "Efficient subgraph matching by postponing cartesian products." Proceedings of the 2016 International Conference on Management of Data. ACM,
25 Supervised Random Walks Predicting and Recommending Links in Social Networks Ronny Pachel Uni Potsdam Feb 07th 2017 Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
26 Contents Introduction 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
27 Link Prediction Introduction Network consisting of nodes and edges Between which two nodes will new edges (links) be created? [In a specific timeframe] Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
28 Challenges Introduction Motivation Author working at Facebook - motivation obvious Every new Link (Friendships, etc.) is increasing Network value Challenges of real Networks Extremely sparse Facebook has nearly 2 billion users, how many friends do you have? To predict that no new Links will be create is nearly 100% accurate! What network features have an influence on the creation of new links? Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
29 Present Work Introduction Method for link prediction and recommendation with supervised random walks Combines network structure and network properties PageRank-like random walk that visits given nodes (positive training examples) more often LEARN edge strength (transition probabilities) Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
30 Contents Supervised Random Walks 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
31 Random Walks Supervised Random Walks Random Walks! 0.4 B A C Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
32 Supervised Random Walks Transition Matrix Probability after one step: = Stationary Probability: /3 1/ /3 = 1/ /3 1/3 Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
33 Overview Supervised Random Walks Given is a graph G Starting from a single node s with already existing Links and for which new links should be predicted Also required is a set of positive and negative training Nodes General approach - can also be used for link prediction, recommendation, link anomaly detection, missing link prediction, etc. Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
34 Approaches Supervised Random Walks We have two general approaches for this problem: As Classification task: Each candidate node is either a created Link in the training data (positive example), or not (negative example) But: Has high class imbalance, which makes learning very hard! How to take the structure of the Network into account? And: The choice and extraction of useful features is very hard! There are countless ways to describe node proximity alone! As Ranking Task: Give Nodes with new created Links a higher Score For Example PageRank and Random Walks with Restarts (RWR) are used Stationary Distribution of a Random Walk gives a measure of closeness Takes the structure into account, but not the properties Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
35 This Approach Supervised Random Walks Combination of Both Approaches - Uses Network Structure and Node and Edge Properties Idea: Control a RWR via its transisiton properties to visit positive training nodes more often than the negative training nodes Trivial Approach by setting the transition properties manually would massively overfit Solution: Find transition properties depending on node and edge properties via a learning algorithm! Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
36 Supervised Random Walks Problem Formulation Given Directed Graph: G(V, E) Start node: s Set of candidate nodes: C = c i = D L Set of destination nodes (positive training examples): D = d i Set of no-link nodes (negative training examples): L = l i Feature Vector for each Edge (u, v): ψ uv Includes the Properties of both, the Nodes and the Edge! Find Edge Strength: a uv Used as transition probability Use Learned Function in Training phase f w (ψ uv ) w: vector of weights Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
37 Supervised Random Walks Predicting Process We take a source node for which we want to predict future links For each Edge (u, v) in G we compute the strength a uv = f w (ψ u v) Then a RWR is run and stationary distribution assigns every node u a probability value p u Nodes are ordered by p u values - highest values are predicted as destinations for future links Weights must be learned! Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
38 Supervised Random Walks Optimization Problem min w F (w) = w 2 + λ d D,l L h(p l p d ) Minimization Problem! p d should be higher than p l h(): loss function - measures how bad the current prediction is in form of a penalty w 2 : second norm of the weight vector - smaller values are wanted! λ: trade-off between the loss and the weights vector A high λ means that the loss is more important than small weights, used to prevent overfitting to the data But how do we come from the input (w) to the measurable output (p)? Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
39 Supervised Random Walks Optimization Problem Set a start value of w For each edge (u,v) calculate the edge transition a uv = f (ψ uv ) Build the random walk transition probability matrix Q Incorporate the restart probabability α into Q Find the vector p of the node probabilities (also known as Personalized PageRank) via a eigenvalue problem (or stationary distribution) This process can be written as a single formula The derivation of this formula is used for a iterative calculation to converge to the p vector (gradient descent method) If I change the weight values like this, is the final result (F(w)) better (smaller)? Note: This problem is not generally convex - no global minimum is guaranteed! Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
40 Contents Experimental Setup 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
41 Network Experimental Setup 4 physics co-authorship networks, complete Facebook network of Iceland Iceland has a very high Facebook penetration Relatively few Edges link to users from other countries Timeframe from Nov 1st 2009 (t 1 ) to Jun 13th 2010 (t 2 ) More than users, about 55% of Icelands population On average 168 friends at t 1 and 26 new friends until t 2 Focus on predicting Links to Nodes that are 2 hops away Empirically most new Links close a triangle Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
42 Preprocessing Experimental Setup Only active Nodes with more than 20 Links Candidate Nodes that have less than 4 common friends have been pruned Empirically shown that it is very unlikely that they would become friends Nodes with too many links have been pruned Even after just 2 hops some nodes can have several millions neighbours These measure speed up the computation but do not influence the prediction performance much Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
43 Learner Setup Experimental Setup From these users, 200 have been randomly selected, 100 as training set to learn the algorithm, and 100 as test set to measure the performance These Properties have been used to generate 7 Features: Edge Age Edge Initiator Communication and Observation Features Number of Common Friends Two performance metrics have been used: AUC - Area under ROC curve Prec@20 - Precision at the Top 20 Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
44 Contents Experiments on Facebook Data 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
45 Algorithm Experiments on Facebook Data We need to choose 4 aspects of the algorithm: Loss Function Edge Strength Function Choice of RWR Restart Parameter α Choice of Regularization Parameter λ Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
46 Experiments on Facebook Data Loss Function Must be continuous and differentiable to optimize over them candidates: Squared loss with margin b: h(x) = max{x + b, 0} 2 Huber Loss with margin b and window z b: 0 if x b, h(x) = (x + b) 2 /(2z) if b < x z b, (x + b) z/2 if x > z b Wilcoxon-Mann-Whitney (WMW) loss with width b: 1 h(x) = 1+exp( x/b) Squared loss and Huber loss can be computed easier WMW has a slightly better performance But: Only WMW increases the AUC and Prec@20! Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
47 Experiments on Facebook Data Strength Function f w (ψ u v) must be non-negative and differentiable Two candidates, both are basing on the inner product of the weight vector and the feature vector candidates: Exponential edge strength: a uv = exp(ψ uv w) Logistic edge strength: a uv = (1 + exp( ψ w)) 1 Experiments show no significant difference in performance But the exponential function has a potential problem with underflowing or overflowing double precision numbers Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
48 Choice of α Experiments on Facebook Data α controls how far a walk wanders before it s reset α = 0 : PageRank of a node is it s degree α = 1 : After one hop the walk is reset Big values are short and local walks, low values are long Have a influence in unweighted graphs, but with edge strengths assigned the influence diminishes Empirically 0.3 < α < 0.7 has same performance Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
49 Choice of λ Experiments on Facebook Data Overfitting is not an issue And number of features is small We chose a λ of 1 Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
50 Results Experiments on Facebook Data SRW with all weights w = 0 as starting point Compared to two unsupervised methods (RWR, node degree) and two supervised machine learning methods (decision trees and logistic regression) Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
51 Contents Conclusion 1 Introduction 2 Supervised Random Walks 3 Experimental Setup 4 Experiments on Facebook Data 5 Conclusion Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
52 Conclusions Conclusion Supervised Random Walks - a new learning algorithm for link prediction and recommendation Combines two different approaches that see the problem either as classification task or as ranking task Demonstrated on real Facebook data Predictions show large improvements over RWR and easier than classical machine learning approaches SRW can be adopted to many other problems Ronny Pachel Uni Potsdam Supervised Random Walks Feb 07th / 28
53 Social Influence Computation and Maximization in Signed Networks with Competing Cascades David Schumann Hasso Plattner Institute Graph Mining Tuesday 7 th February, / 12
54 What is this paper about? Domain: Network of susceptible nodes. Trust and Distrust Relationships Diffusion Model: Independent Cascade Problem: Starting seed-set of size m with the highest expectation of nodes colored red Signed Network Influence Maximization (SiNiMax) Problem NP-hard 2 / 12
55 What is this paper about? 3 / 12
56 Signed Networks with competing Cascades Weighted, directed and signed graph G = (V, E, W ) Weight Matrix W : sign represents trust or distrust relationships. Absolute value denotes strength of influence / 12
57 Signed Networks with competing Cascades Two influnce diffusion processes spreading in discrete time steps Using the standard Independent Cascade Model (ICM) Infected node v gets one chance to infect neighbour w with probability p v,w On success w becomes infected at step t + 1. Infected nodes never return to susceptible state / 12
58 Signed Networks with competing Cascades Two influnce diffusion processes spreading in discrete time steps Using the standard Independent Cascade Model (ICM) Infected node v gets one chance to infect neighbour w with probability p v,w On success w becomes infected at step t + 1. Infected nodes never return to susceptible state / 12
59 Signed Networks with competing Cascades Two influnce diffusion processes spreading in discrete time steps Using the standard Independent Cascade Model (ICM) Infected node v gets one chance to infect neighbour w with probability p v,w On success w becomes infected at step t + 1. Infected nodes never return to susceptible state / 12
60 Signed Networks with competing Cascades Two influnce diffusion processes spreading in discrete time steps Using the standard Independent Cascade Model (ICM) Infected node v gets one chance to infect neighbour w with probability p v,w On success w becomes infected at step t + 1. Infected nodes never return to susceptible state / 12
61 Help of the Unified Model Reduces computational cost for commonly used diffusion models. Using: Individual and Collective Influence. For ICM: p (v,u) (t) = 0 r u,t = 1 P (no infected neighbor succeeds) r u,t = 1 (1 p(b v,t 1 ))... v N i (u) 6 / 12
62 Unified Model of Competing Cascades r + u,t r u,t B + u,t B u,t = Collective influence on u to be red. = Collective influence on u to be blue. = Probability of u being red at or before time t. = Probability of u being blue at or before time t. B u,t + (k, +) = Probability of node u being colored red at time t if node k is infected with red at time t 1. 7 / 12
63 Online Seed-set Selection using the Unified Model for Signed Networks: OSSU M ± Algorithm 1 pseudocode for OSSUM± Input: G, m Output: S 1: S 2: for t = 1 to m do 3: (k, c) = arg max (k,c) u (B+ u,t (k, c) B+ u,t ) 4: S S (k, c) 5: end for Runtime Complexity: O(m( V + E )) 8 / 12
64 Experiment Results Comparison of OSSU M ± to three other heuristics: Positive Degree: Nodes with highest positive degree are selected Positive Degree Discount: Variation of Positive Degree Effective Degree: Nodes with highest effective degree are selected Data Set: Epinions Slashdot # nodes 131k 82k # edges 840k 550k # positive edges 710k 425k # negative edges 130k 125k 9 / 12
65 Experiment Results 10 / 12
66 Experiment Results 11 / 12
67 Review Positive: Introduced novel influence maximization problem SiNiMax OSSU M ± outperforms state-of-the-art heuristics for influence maximization Application to real world data Negative: Inaccuracies in description of method (missing edge cases) Unproven claim of extensaibility of this approach Real world data had to be modified to show advantage of OSSUM± 12 / 12
68 Top-k Link Recommendation in Social Graphs Sören Tietböhl
69
70
71
72
73 X = UV
74 X = UV
75
76 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6)
77 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases
78 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List:
79 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List: 1. Define new loss function that penalizes mistakes at the top more than at the bottom 3 4 6
80 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List: 1. Define new loss function that penalizes mistakes at the top more than at the bottom 2. Derive an algorithm that minimizes that function 3 4 6
81 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List: 1. Define new loss function that penalizes mistakes at the top more than at the bottom 2. Derive an algorithm that minimizes that function 3. Evaluate the algorithm using various datasets and metrics 3 4 6
82 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Loss is 1 in both cases TODO List: 1. Define new loss function that penalizes mistakes at the top more than at the bottom 2. Derive an algorithm that minimizes that function 3. Evaluate the algorithm using various datasets and metrics 4.??? 5. Profit
83
84
85 = Training Adjacency Matrix = Ranking Score
86 = Training Adjacency Matrix = Ranking Score (2,4),(4,5),(3,4),(1,4),(4,6)
87 = Training Adjacency Matrix = Ranking Score (2,4),(4,5),(3,4),(1,4),(4,6)
88 = Training Adjacency Matrix = Ranking Score (2,4),(4,5),(3,4),(1,4),(4,6)
89 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6)
90 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6)
91 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Prediction 1:
92 Predictions for user 4: Ground truth: (1,4),(2,4),(3,4),(4,5),(4,6) Example prediction 1: (2,4),(3,4),(4,5),(1,4),(4,6) Example prediction 2: (2,4),(4,5),(3,4),(1,4),(4,6) Prediction 1: Prediction 2:
93
94
95 Sigmoid function:
96 Sigmoid function: Now its derivable!
97 Sigmoid function: Now its derivable! yay
98 Sigmoid function: Now its derivable!
99
100
101
102
103
104 (2,4),(3,4),(4,5),(1,4),(4,6)
105 (2,4),(3,4),(4,5),(1,4),(4,6) (1,4),(2,4),(3,4),(4,5),(4,6)
106
107
108
109
110 Sources Dongjin Song, David A. Meyer, Dacheng Tao, "Top-k Link Recommendation in Social Networks", 2015 IEEE International Conference on Data Mining (ICDM), vol. 00, no., pp , 2015, doi: /icdm Graph Mining Course Slides quicklatex.com
111 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel Graph Mining WT 2016/17 1 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
112 Structural Graph Clustering Cluster v 1 v 9 Hub v 2 v 4 v 5 v 6 v 8 Outlier v 3 v 7 v 10 2 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
113 Agenda Principles of SCAN Improvements of pscan Experimental Results 3 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
114 Principles of SCAN Graph G = (V, E) Structural Neighborhood: N u = {v V (u, v) E} {u} d u = N u Vertex u v 1 v 9 Neighborhood N u v 2 v 4 v 5 v 6 v 8 v 3 v 7 v 10 4 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
115 Principles of SCAN Structural Similarity: σ(u, v) = N u N v du d v Neighborhood N u v 1 v 9 Neighborhood N v v 2 v 4 v 5 v 6 v 8 N u N v v 3 v 7 σ(v 6, v 5 ) = v 10 5 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
116 Principles of SCAN Similarity Threshold and Core Vertices: Nu ɛ = {v N u σ(u, v) ɛ} V C = {v V Nv ɛ µ} Core Vertex ɛ-edge ɛ = 0.7 v v v v v v v 9 v v 8 µ = 4 v 10 6 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
117 Principles of SCAN Structure Reachability: u is structure reachable by v if there is a Path of ɛ-edges and intermediate core vertices Cluster C V such that all Vertices v C are mutually structure reachable and no Vertex x V \ C is structure reachable by any Vertex in C Core Vertex v 1 v 9 ɛ-edge v 2 v 4 v 5 v 6 v 8 ɛ = 0.7 v 3 v 7 µ = 4 v 10 7 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
118 Principles of SCAN Algorithm: SCAN requires structural similarity calculation along all edges in G equivalent to enumeration of all triangles in G complexity of O( E 1.5 ) which is worst case optimum 8 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
119 Agenda Principles of SCAN Improvements of pscan Experimental Results 9 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
120 Improvements of pscan Paradigm: 1 cluster core vertices 2 expand clusters by core vertices structural neighborhood Effective Degree and Similar Degree: vertex u V with incomplete similarity measures in its neighborhood effective degree: d e u = d u x where x is the number of neighbors v V with σ(u, v) < ɛ known similar degree: d s u number of neighbors v V with σ(u, v) ɛ known d e u and d s u provide an upper and lower bound for N ɛ u calculate structural neighborhoods for vertices in non-increasing d e -order update d e and d s for every calculated σ; omit neighborhood calculation if d e or d s are sufficient to identify core vertex 10 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
121 Improvements of pscan u ds u d e u v v v v v v v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
122 Improvements of pscan u ds u d e u v v v v 4 5 v 5 3 v 6 4 v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
123 Improvements of pscan u ds u d e u v v v v 4 4 v 5 2 v 6 4 v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
124 Improvements of pscan u ds u d e u v v v v 4 4 v 5 2 v v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
125 Improvements of pscan u ds u d e u v v v v 4 4 v 5 2 v v ɛ = 0.7 v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
126 Improvements of pscan u ds u d e u 1 v v ɛ = 0.7 v v v 4 4 v 5 2 v v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
127 Improvements of pscan u ds u d e u 1 v v ɛ = 0.7 v v v 4 4 v 5 2 v v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
128 Improvements of pscan u ds u d e u 1 v v ɛ = 0.7 v v v 4 4 v 5 2 v v v µ = 4 v pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
129 Agenda Principles of SCAN Improvements of pscan Experimental Results 19 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
130 Experimental Results Algorithms: SCAN SCAN++ (approximative version of SCAN omitting some similarity calculations) pscan (5 variants with different degrees of optimization) Data: 13 real world networks (13k to 1.2G edges) 11 generated networks with varying characteristics 20 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
131 Experimental Results 1 order of magnitude faster than either SCAN or SCAN++ 21 pscan: Fast and Exact Structural Graph Clustering Lukas Wenzel WT 2016/17
132 Clustering Large Probabilistic Graphs Marianne Thieffry
133 Agenda 1. Definition of pclusteredit 2. Algorithms for pclusteredit 3. Generalizations of pclusteredit
134 Definition of pclusteredit
135 Probabilistic Graphs
136 Probabilistic Graphs
137 Probabilistic Graphs
138 Cluster graphs
139 Edit distance # edges added + # edges removed
140 Edit distance # edges added + # edges removed [between all deterministic instances of a given probabilistic graph and a given deterministic graph ]
141 pclusteredit [while minimizing the edit distance]
142 pclusteredit
143 CorrelationClustering
144 CorrelationClustering
145 CorrelationClustering
146 Algorithms for pclusteredit
147 pkwikcluster
148 pkwikcluster
149 pkwikcluster
150 pkwikcluster
151 pkwikcluster 0.8
152 pkwikcluster O(m)
153 Furthest
154 Furthest
155 Furthest
156 Furthest
157 Furthest O(nk²)
158 Agglomerative Singleton clusters Merge clusters with highest average edge probability, until highest average edge probability is at most 0.5 Complexity O(kn log n)
159 pkwikcluster
160 pkwikcluster
161 pkwikcluster
162 pkwikcluster
163 pkwikcluster O(kn log n)
164 Significance
165 Evaluation - CORE dataset protein-protein interaction network 2708 nodes 7123 edges
166 Evaluation - CORE dataset
167 Evaluation - CORE dataset
168 Evaluation - CORE dataset
169 Evaluation - CORE dataset
170 Generalizations of pclusteredit
171 Noisy cluster graphs
172 p2clusteredit & minα Find α-cluster with minimal edit distance Find α that produces minimum α-cluster with minimal edit distance
173 The shortest Path is not always a straight line From Vasiliki Kalavri, Tiago Simas & Dionysios Logothetis Philippe ZOU Graph Mining - WS 2016/2017 Davide Mottin & Konstantina Lazaridou 1
174 Agenda 1. Overview 2. Implementation 3. Evaluation 4. Conclusion 2
175 Overview 3
176 Motivations Leverage the Backbone metric The Blackbone metric is the minimum subgraph that preserves the shortest paths of a weighted graph Improve graph analysis tools and computing performance Large graph reduce running time and storage 4
177 Backbone Metric For instance, in the example they give : 5
178 Concept of semi-metricity An edge is semi-metric if you can find an indirect shorter path between the edges. Let consider the previous graph 6
179 Concept of semi-metricity An edge is semi-metric if you can find an indirect shorter path between the edges. Let consider the previous graph = 6 < 10 AD is semi-metric (same for the node CE) 7
180 Concept of semi-metricity Semi-metricity Degree : number of edges to add to reach the shortest path compared to the original (straight) one AD will be 2nd order (and CE will be 1st order) 8
181 Concept of semi-metricity Performed an analysis on data sets 9
182 Implementation 10
183 Backbone algorithm 3 phases Find 1st order semi-metric edges Only look at triangles 11
184 Backbone algorithm 3 phases Find 1st order semi-metric edges Only look at triangles 12
185 Backbone algorithm 3 phases Focus on 2nd order semi-metric edges 13
186 Backbone algorithm 3 phases Focus on 2nd order semi-metric edges Lowest weight edge of every node is metric? 14
187 Backbone algorithm 3 phases Breadth first search? 15
188 Backbone algorithm 3 phases Breadth first search? 16
189 Backbone algorithm 3 phases Breadth first search? 17
190 Backbone algorithm 3 phases Breadth first search? Target reached but with a shorter path -> Semi-metric edge 18
191 Backbone algorithm 3 phases End of algorithm 19
192 Evaluation 20
193 Execution time for phase 1 Execution time for finding triangles + removal 1st order semi-metric 21
194 Execution time for phase 1 Increase linearly. 15 Minutes for 4 Billions edges. 22
195 Storage Size reduction of the network after removal 23
196 Conclusion 24
197 Conclusion Provide a new concept in graphs : semi-metricity Algorithm to lower the number of edges in order to improve computation and reduce time show that direct nodes are not necessarily the shortest ones Challenge the large graphes issue 25
Link Prediction. Eman Badr Mohammed Saquib Akmal Khan
Link Prediction Eman Badr Mohammed Saquib Akmal Khan 11-06-2013 Link Prediction Which pair of nodes should be connected? Applications Facebook friend suggestion Recommendation systems Monitoring and controlling
More informationOverlapping Community Detection at Scale: A Nonnegative Matrix Factorization Approach
Overlapping Community Detection at Scale: A Nonnegative Matrix Factorization Approach Author: Jaewon Yang, Jure Leskovec 1 1 Venue: WSDM 2013 Presenter: Yupeng Gu 1 Stanford University 1 Background Community
More informationLogistic Regression Logistic
Case Study 1: Estimating Click Probabilities L2 Regularization for Logistic Regression Machine Learning/Statistics for Big Data CSE599C1/STAT592, University of Washington Carlos Guestrin January 10 th,
More informationMore on NP and Reductions
Indian Institute of Information Technology Design and Manufacturing, Kancheepuram Chennai 600 127, India An Autonomous Institute under MHRD, Govt of India http://www.iiitdm.ac.in COM 501 Advanced Data
More informationComputing Competing Cascades on Signed Networks
Noname manuscript No. (will be inserted by the editor Computing Competing Cascades on Signed Networks Ajitesh Srivastava Charalampos Chelmis Viktor K. Prasanna Received: date / Accepted: date Abstract
More informationDetermining the Diameter of Small World Networks
Determining the Diameter of Small World Networks Frank W. Takes & Walter A. Kosters Leiden University, The Netherlands CIKM 2011 October 2, 2011 Glasgow, UK NWO COMPASS project (grant #12.0.92) 1 / 30
More informationCSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18
CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$
More informationTufts COMP 135: Introduction to Machine Learning
Tufts COMP 135: Introduction to Machine Learning https://www.cs.tufts.edu/comp/135/2019s/ Logistic Regression Many slides attributable to: Prof. Mike Hughes Erik Sudderth (UCI) Finale Doshi-Velez (Harvard)
More informationWeb Structure Mining Nodes, Links and Influence
Web Structure Mining Nodes, Links and Influence 1 Outline 1. Importance of nodes 1. Centrality 2. Prestige 3. Page Rank 4. Hubs and Authority 5. Metrics comparison 2. Link analysis 3. Influence model 1.
More informationCSCI-567: Machine Learning (Spring 2019)
CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Mar. 19, 2019 March 19, 2019 1 / 43 Administration March 19, 2019 2 / 43 Administration TA3 is due this week March
More informationPredictive analysis on Multivariate, Time Series datasets using Shapelets
1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,
More informationIntroduction to Logistic Regression
Introduction to Logistic Regression Guy Lebanon Binary Classification Binary classification is the most basic task in machine learning, and yet the most frequent. Binary classifiers often serve as the
More informationGreedy Maximization Framework for Graph-based Influence Functions
Greedy Maximization Framework for Graph-based Influence Functions Edith Cohen Google Research Tel Aviv University HotWeb '16 1 Large Graphs Model relations/interactions (edges) between entities (nodes)
More informationClustering. CSL465/603 - Fall 2016 Narayanan C Krishnan
Clustering CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Supervised vs Unsupervised Learning Supervised learning Given x ", y " "%& ', learn a function f: X Y Categorical output classification
More informationCS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine
CS 277: Data Mining Mining Web Link Structure Class Presentations In-class, Tuesday and Thursday next week 2-person teams: 6 minutes, up to 6 slides, 3 minutes/slides each person 1-person teams 4 minutes,
More informationProtein Complex Identification by Supervised Graph Clustering
Protein Complex Identification by Supervised Graph Clustering Yanjun Qi 1, Fernanda Balem 2, Christos Faloutsos 1, Judith Klein- Seetharaman 1,2, Ziv Bar-Joseph 1 1 School of Computer Science, Carnegie
More informationAd Placement Strategies
Case Study : Estimating Click Probabilities Intro Logistic Regression Gradient Descent + SGD AdaGrad Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox January 7 th, 04 Ad
More informationOnline Social Networks and Media. Link Analysis and Web Search
Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information
More informationOverlapping Communities
Overlapping Communities Davide Mottin HassoPlattner Institute Graph Mining course Winter Semester 2017 Acknowledgements Most of this lecture is taken from: http://web.stanford.edu/class/cs224w/slides GRAPH
More informationCase Study 1: Estimating Click Probabilities. Kakade Announcements: Project Proposals: due this Friday!
Case Study 1: Estimating Click Probabilities Intro Logistic Regression Gradient Descent + SGD Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade April 4, 017 1 Announcements:
More informationSummary and discussion of: Dropout Training as Adaptive Regularization
Summary and discussion of: Dropout Training as Adaptive Regularization Statistics Journal Club, 36-825 Kirstin Early and Calvin Murdock November 21, 2014 1 Introduction Multi-layered (i.e. deep) artificial
More informationSVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning
SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning Mark Schmidt University of British Columbia, May 2016 www.cs.ubc.ca/~schmidtm/svan16 Some images from this lecture are
More informationMachine Learning for NLP
Machine Learning for NLP Linear Models Joakim Nivre Uppsala University Department of Linguistics and Philology Slides adapted from Ryan McDonald, Google Research Machine Learning for NLP 1(26) Outline
More informationCommunities Via Laplacian Matrices. Degree, Adjacency, and Laplacian Matrices Eigenvectors of Laplacian Matrices
Communities Via Laplacian Matrices Degree, Adjacency, and Laplacian Matrices Eigenvectors of Laplacian Matrices The Laplacian Approach As with betweenness approach, we want to divide a social graph into
More informationCPSC 340: Machine Learning and Data Mining
CPSC 340: Machine Learning and Data Mining Linear Classifiers: predictions Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due Friday of next
More informationTraining the linear classifier
215, Training the linear classifier A natural way to train the classifier is to minimize the number of classification errors on the training data, i.e. choosing w so that the training error is minimized.
More informationGradient Boosting (Continued)
Gradient Boosting (Continued) David Rosenberg New York University April 4, 2016 David Rosenberg (New York University) DS-GA 1003 April 4, 2016 1 / 31 Boosting Fits an Additive Model Boosting Fits an Additive
More informationDATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS
DATA MINING LECTURE 3 Link Analysis Ranking PageRank -- Random walks HITS How to organize the web First try: Manually curated Web Directories How to organize the web Second try: Web Search Information
More informationData Mining and Analysis: Fundamental Concepts and Algorithms
Data Mining and Analysis: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More informationLinear classifiers: Overfitting and regularization
Linear classifiers: Overfitting and regularization Emily Fox University of Washington January 25, 2017 Logistic regression recap 1 . Thus far, we focused on decision boundaries Score(x i ) = w 0 h 0 (x
More information18.9 SUPPORT VECTOR MACHINES
744 Chapter 8. Learning from Examples is the fact that each regression problem will be easier to solve, because it involves only the examples with nonzero weight the examples whose kernels overlap the
More informationMachine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring /
Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring 2015 http://ce.sharif.edu/courses/93-94/2/ce717-1 / Agenda Combining Classifiers Empirical view Theoretical
More informationMidterm Exam, Spring 2005
10-701 Midterm Exam, Spring 2005 1. Write your name and your email address below. Name: Email address: 2. There should be 15 numbered pages in this exam (including this cover sheet). 3. Write your name
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table
More informationA New Space for Comparing Graphs
A New Space for Comparing Graphs Anshumali Shrivastava and Ping Li Cornell University and Rutgers University August 18th 2014 Anshumali Shrivastava and Ping Li ASONAM 2014 August 18th 2014 1 / 38 Main
More informationOverfitting, Bias / Variance Analysis
Overfitting, Bias / Variance Analysis Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 8, 207 / 40 Outline Administration 2 Review of last lecture 3 Basic
More informationOnline Social Networks and Media. Link Analysis and Web Search
Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information
More informationCapacity Releasing Diffusion for Speed and Locality
000 00 002 003 004 005 006 007 008 009 00 0 02 03 04 05 06 07 08 09 020 02 022 023 024 025 026 027 028 029 030 03 032 033 034 035 036 037 038 039 040 04 042 043 044 045 046 047 048 049 050 05 052 053 054
More informationThe exam is closed book, closed notes except your one-page cheat sheet.
CS 189 Fall 2015 Introduction to Machine Learning Final Please do not turn over the page before you are instructed to do so. You have 2 hours and 50 minutes. Please write your initials on the top-right
More informationUnsupervised Learning
2018 EE448, Big Data Mining, Lecture 7 Unsupervised Learning Weinan Zhang Shanghai Jiao Tong University http://wnzhang.net http://wnzhang.net/teaching/ee448/index.html ML Problem Setting First build and
More informationMax Margin-Classifier
Max Margin-Classifier Oliver Schulte - CMPT 726 Bishop PRML Ch. 7 Outline Maximum Margin Criterion Math Maximizing the Margin Non-Separable Data Kernels and Non-linear Mappings Where does the maximization
More informationKristina Lerman USC Information Sciences Institute
Rethinking Network Structure Kristina Lerman USC Information Sciences Institute Università della Svizzera Italiana, December 16, 2011 Measuring network structure Central nodes Community structure Strength
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu March 16, 2016 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision
More informationSVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels
SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels Karl Stratos June 21, 2018 1 / 33 Tangent: Some Loose Ends in Logistic Regression Polynomial feature expansion in logistic
More informationData Mining Techniques
Data Mining Techniques CS 622 - Section 2 - Spring 27 Pre-final Review Jan-Willem van de Meent Feedback Feedback https://goo.gl/er7eo8 (also posted on Piazza) Also, please fill out your TRACE evaluations!
More informationLearning theory. Ensemble methods. Boosting. Boosting: history
Learning theory Probability distribution P over X {0, 1}; let (X, Y ) P. We get S := {(x i, y i )} n i=1, an iid sample from P. Ensemble methods Goal: Fix ɛ, δ (0, 1). With probability at least 1 δ (over
More informationOnline Advertising is Big Business
Online Advertising Online Advertising is Big Business Multiple billion dollar industry $43B in 2013 in USA, 17% increase over 2012 [PWC, Internet Advertising Bureau, April 2013] Higher revenue in USA
More informationRegression with Numerical Optimization. Logistic
CSG220 Machine Learning Fall 2008 Regression with Numerical Optimization. Logistic regression Regression with Numerical Optimization. Logistic regression based on a document by Andrew Ng October 3, 204
More informationPoint-of-Interest Recommendations: Learning Potential Check-ins from Friends
Point-of-Interest Recommendations: Learning Potential Check-ins from Friends Huayu Li, Yong Ge +, Richang Hong, Hengshu Zhu University of North Carolina at Charlotte + University of Arizona Hefei University
More informationResearch Overview. Kristjan Greenewald. February 2, University of Michigan - Ann Arbor
Research Overview Kristjan Greenewald University of Michigan - Ann Arbor February 2, 2016 2/17 Background and Motivation Want efficient statistical modeling of high-dimensional spatio-temporal data with
More informationMark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.
CS 189 Spring 2015 Introduction to Machine Learning Midterm You have 80 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. No calculators or electronic items.
More informationClassification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012
Classification CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Topics Discriminant functions Logistic regression Perceptron Generative models Generative vs. discriminative
More informationSupport Vector Machines
Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized
More informationECEN 689 Special Topics in Data Science for Communications Networks
ECEN 689 Special Topics in Data Science for Communications Networks Nick Duffield Department of Electrical & Computer Engineering Texas A&M University Lecture 8 Random Walks, Matrices and PageRank Graphs
More informationSUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION
SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology
More informationUNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013
UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 Exam policy: This exam allows two one-page, two-sided cheat sheets; No other materials. Time: 2 hours. Be sure to write your name and
More informationMachine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.
Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted
More informationLinear and Logistic Regression. Dr. Xiaowei Huang
Linear and Logistic Regression Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Two Classical Machine Learning Algorithms Decision tree learning K-nearest neighbor Model Evaluation Metrics
More informationCollaborative Filtering. Radek Pelánek
Collaborative Filtering Radek Pelánek 2017 Notes on Lecture the most technical lecture of the course includes some scary looking math, but typically with intuitive interpretation use of standard machine
More informationCS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu 10/24/2012 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu
More informationDecision Trees: Overfitting
Decision Trees: Overfitting Emily Fox University of Washington January 30, 2017 Decision tree recap Loan status: Root 22 18 poor 4 14 Credit? Income? excellent 9 0 3 years 0 4 Fair 9 4 Term? 5 years 9
More informationSpectral Methods for Subgraph Detection
Spectral Methods for Subgraph Detection Nadya T. Bliss & Benjamin A. Miller Embedded and High Performance Computing Patrick J. Wolfe Statistics and Information Laboratory Harvard University 12 July 2010
More informationSubmodular Functions Properties Algorithms Machine Learning
Submodular Functions Properties Algorithms Machine Learning Rémi Gilleron Inria Lille - Nord Europe & LIFL & Univ Lille Jan. 12 revised Aug. 14 Rémi Gilleron (Mostrare) Submodular Functions Jan. 12 revised
More informationPATTERN CLASSIFICATION
PATTERN CLASSIFICATION Second Edition Richard O. Duda Peter E. Hart David G. Stork A Wiley-lnterscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane Singapore Toronto CONTENTS
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationLearning Tetris. 1 Tetris. February 3, 2009
Learning Tetris Matt Zucker Andrew Maas February 3, 2009 1 Tetris The Tetris game has been used as a benchmark for Machine Learning tasks because its large state space (over 2 200 cell configurations are
More informationPerformance Evaluation. Analyzing competitive influence maximization problems with partial information: An approximation algorithmic framework
Performance Evaluation ( ) Contents lists available at ScienceDirect Performance Evaluation journal homepage: www.elsevier.com/locate/peva Analyzing competitive influence maximization problems with partial
More informationPattern Recognition Approaches to Solving Combinatorial Problems in Free Groups
Contemporary Mathematics Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups Robert M. Haralick, Alex D. Miasnikov, and Alexei G. Myasnikov Abstract. We review some basic methodologies
More informationUsing R for Iterative and Incremental Processing
Using R for Iterative and Incremental Processing Shivaram Venkataraman, Indrajit Roy, Alvin AuYoung, Robert Schreiber UC Berkeley and HP Labs UC BERKELEY Big Data, Complex Algorithms PageRank (Dominant
More informationFinal Overview. Introduction to ML. Marek Petrik 4/25/2017
Final Overview Introduction to ML Marek Petrik 4/25/2017 This Course: Introduction to Machine Learning Build a foundation for practice and research in ML Basic machine learning concepts: max likelihood,
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu November 16, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision
More informationApprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning
Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning Nicolas Thome Prenom.Nom@cnam.fr http://cedric.cnam.fr/vertigo/cours/ml2/ Département Informatique Conservatoire
More informationAnomaly Detection. Davide Mottin, Konstantina Lazaridou. HassoPlattner Institute. Graph Mining course Winter Semester 2016
Anomaly Detection Davide Mottin, Konstantina Lazaridou HassoPlattner Institute Graph Mining course Winter Semester 2016 and Next week February 7, third round presentations Slides are due by February 6,
More informationAnalytically tractable processes on networks
University of California San Diego CERTH, 25 May 2011 Outline Motivation 1 Motivation Networks Random walk and Consensus Epidemic models Spreading processes on networks 2 Networks Motivation Networks Random
More informationCPSC 340: Machine Learning and Data Mining. Gradient Descent Fall 2016
CPSC 340: Machine Learning and Data Mining Gradient Descent Fall 2016 Admin Assignment 1: Marks up this weekend on UBC Connect. Assignment 2: 3 late days to hand it in Monday. Assignment 3: Due Wednesday
More informationStatistical Learning. Dong Liu. Dept. EEIS, USTC
Statistical Learning Dong Liu Dept. EEIS, USTC Chapter 6. Unsupervised and Semi-Supervised Learning 1. Unsupervised learning 2. k-means 3. Gaussian mixture model 4. Other approaches to clustering 5. Principle
More informationAnomaly Detection for the CERN Large Hadron Collider injection magnets
Anomaly Detection for the CERN Large Hadron Collider injection magnets Armin Halilovic KU Leuven - Department of Computer Science In cooperation with CERN 2018-07-27 0 Outline 1 Context 2 Data 3 Preprocessing
More informationLogistic Regression & Neural Networks
Logistic Regression & Neural Networks CMSC 723 / LING 723 / INST 725 Marine Carpuat Slides credit: Graham Neubig, Jacob Eisenstein Logistic Regression Perceptron & Probabilities What if we want a probability
More informationEmpirical Risk Minimization
Empirical Risk Minimization Fabrice Rossi SAMM Université Paris 1 Panthéon Sorbonne 2018 Outline Introduction PAC learning ERM in practice 2 General setting Data X the input space and Y the output space
More informationIs the test error unbiased for these programs?
Is the test error unbiased for these programs? Xtrain avg N o Preprocessing by de meaning using whole TEST set 2017 Kevin Jamieson 1 Is the test error unbiased for this program? e Stott see non for f x
More informationLearning from Examples
Learning from Examples Data fitting Decision trees Cross validation Computational learning theory Linear classifiers Neural networks Nonparametric methods: nearest neighbor Support vector machines Ensemble
More informationThe exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet.
CS 189 Spring 013 Introduction to Machine Learning Final You have 3 hours for the exam. The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet. Please
More informationDATA MINING AND MACHINE LEARNING
DATA MINING AND MACHINE LEARNING Lecture 5: Regularization and loss functions Lecturer: Simone Scardapane Academic Year 2016/2017 Table of contents Loss functions Loss functions for regression problems
More informationhsnim: Hyper Scalable Network Inference Machine for Scale-Free Protein-Protein Interaction Networks Inference
CS 229 Project Report (TR# MSB2010) Submitted 12/10/2010 hsnim: Hyper Scalable Network Inference Machine for Scale-Free Protein-Protein Interaction Networks Inference Muhammad Shoaib Sehgal Computer Science
More informationClustering. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms March 8, / 26
Clustering Professor Ameet Talwalkar Professor Ameet Talwalkar CS26 Machine Learning Algorithms March 8, 217 1 / 26 Outline 1 Administration 2 Review of last lecture 3 Clustering Professor Ameet Talwalkar
More informationVariations of Logistic Regression with Stochastic Gradient Descent
Variations of Logistic Regression with Stochastic Gradient Descent Panqu Wang(pawang@ucsd.edu) Phuc Xuan Nguyen(pxn002@ucsd.edu) January 26, 2012 Abstract In this paper, we extend the traditional logistic
More informationMulticlass Classification-1
CS 446 Machine Learning Fall 2016 Oct 27, 2016 Multiclass Classification Professor: Dan Roth Scribe: C. Cheng Overview Binary to multiclass Multiclass SVM Constraint classification 1 Introduction Multiclass
More informationSupport vector machines Lecture 4
Support vector machines Lecture 4 David Sontag New York University Slides adapted from Luke Zettlemoyer, Vibhav Gogate, and Carlos Guestrin Q: What does the Perceptron mistake bound tell us? Theorem: The
More informationData Mining Part 5. Prediction
Data Mining Part 5. Prediction 5.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline How the Brain Works Artificial Neural Networks Simple Computing Elements Feed-Forward Networks Perceptrons (Single-layer,
More informationUsing Local Spectral Methods in Theory and in Practice
Using Local Spectral Methods in Theory and in Practice Michael W. Mahoney ICSI and Dept of Statistics, UC Berkeley ( For more info, see: http: // www. stat. berkeley. edu/ ~ mmahoney or Google on Michael
More informationA Gradient-based Adaptive Learning Framework for Efficient Personal Recommendation
A Gradient-based Adaptive Learning Framework for Efficient Personal Recommendation Yue Ning 1 Yue Shi 2 Liangjie Hong 2 Huzefa Rangwala 3 Naren Ramakrishnan 1 1 Virginia Tech 2 Yahoo Research. Yue Shi
More informationNearest Neighbor. Machine Learning CSE546 Kevin Jamieson University of Washington. October 26, Kevin Jamieson 2
Nearest Neighbor Machine Learning CSE546 Kevin Jamieson University of Washington October 26, 2017 2017 Kevin Jamieson 2 Some data, Bayes Classifier Training data: True label: +1 True label: -1 Optimal
More information6.207/14.15: Networks Lecture 7: Search on Networks: Navigation and Web Search
6.207/14.15: Networks Lecture 7: Search on Networks: Navigation and Web Search Daron Acemoglu and Asu Ozdaglar MIT September 30, 2009 1 Networks: Lecture 7 Outline Navigation (or decentralized search)
More information2018 EE448, Big Data Mining, Lecture 4. (Part I) Weinan Zhang Shanghai Jiao Tong University
2018 EE448, Big Data Mining, Lecture 4 Supervised Learning (Part I) Weinan Zhang Shanghai Jiao Tong University http://wnzhang.net http://wnzhang.net/teaching/ee448/index.html Content of Supervised Learning
More informationGenerative v. Discriminative classifiers Intuition
Logistic Regression (Continued) Generative v. Discriminative Decision rees Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University January 31 st, 2007 2005-2007 Carlos Guestrin 1 Generative
More informationIntroduction to Machine Learning. Introduction to ML - TAU 2016/7 1
Introduction to Machine Learning Introduction to ML - TAU 2016/7 1 Course Administration Lecturers: Amir Globerson (gamir@post.tau.ac.il) Yishay Mansour (Mansour@tau.ac.il) Teaching Assistance: Regev Schweiger
More informationA Randomized Approach for Crowdsourcing in the Presence of Multiple Views
A Randomized Approach for Crowdsourcing in the Presence of Multiple Views Presenter: Yao Zhou joint work with: Jingrui He - 1 - Roadmap Motivation Proposed framework: M2VW Experimental results Conclusion
More informationMODULE -4 BAYEIAN LEARNING
MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities
More informationMaximizing the Spread of Influence through a Social Network. David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03
Maximizing the Spread of Influence through a Social Network David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03 Influence and Social Networks Economics, sociology, political science, etc. all have studied
More information