Faloutsos, Kolda, Sun

Size: px
Start display at page:

Download "Faloutsos, Kolda, Sun"

Transcription

1 Faloutsos, Kolda, Sun SDM'7 About the tutorial Mining Large Time-evolving Data Using Matrix and Tensor Tools Christos Faloutsos Carnegie Mellon Univ. Tamara G. Kolda Sandia National Labs Jimeng Sun Carnegie Mellon Univ. Introduce matrix and tensor tools through real mining applications Goal: find patterns, rules, clusters, outliers, in matrices and in tensors SDM'7 Faloutsos, Kolda, Sun 1-2 Motivation 1: Why matrix? Why matrices are important? John Peter Mary Nick... Examples of Matrices: Graph - social network John Peter Mary Nick SDM'7 Faloutsos, Kolda, Sun 1-3 SDM'7 Faloutsos, Kolda, Sun 1-4 Examples of Matrices: cloud of n-d points Examples of Matrices: Market basket market basket as in Association Rules John Peter Mary Nick... chol# blood# age John Peter Mary Nick... milk bread choc. wine SDM'7 Faloutsos, Kolda, Sun 1-5 SDM'7 Faloutsos, Kolda, Sun 1-6 1

2 Faloutsos, Kolda, Sun SDM'7 Examples of Matrices: Documents and terms Examples of Matrices: Authors and terms Paper#1 Paper#2 Paper#3 Paper#4... data mining classif. tree John Peter Mary Nick... data mining classif. tree SDM'7 Faloutsos, Kolda, Sun 1-7 SDM'7 Faloutsos, Kolda, Sun 1-8 Examples of Matrices: sensor-ids and time-ticks Motivation 2: Why tensor? Q: what is a tensor? t1 t2 t3 t4... temp1 temp2 humid. pressure SDM'7 Faloutsos, Kolda, Sun 1-9 SDM'7 Faloutsos, Kolda, Sun 1-1 Motivation 2: Why tensor? A: N-D generalization of matrix: Motivation 2: Why tensor? A: N-D generalization of matrix: SDM 7 John Peter Mary Nick... data mining classif. tree John Peter Mary Nick... SDM 5 SDM 6 SDM 7 data mining classif. tree SDM'7 Faloutsos, Kolda, Sun 1-11 SDM'7 Faloutsos, Kolda, Sun

3 Faloutsos, Kolda, Sun SDM'7 Mode#2 Tensors are useful for 3 or more modes Terminology: mode (or aspect ): Mode#3 data mining classif. tree Mode ( aspect) #1 SDM'7 Faloutsos, Kolda, Sun 1-13 Motivating Applications Why matrices are important? Why tensors are useful? P1: environmental sensors P2: data center monitoring ( autonomic ) P3: social networks P4: network forensics P5: web mining SDM'7 Faloutsos, Kolda, Sun P1: Environmental sensor monitoring 6 P2: Clusters/data center monitoring value time (min) Temperature Data in three aspects (time, location, type) value time (min) Light value time (min) time (min) Humidity Voltage SDM'7 Faloutsos, Kolda, Sun 1-15 value Monitor correlations of multiple measurements Automatically flag anomalous behavior Intemon: intelligent monitoring system Prof. Greg Ganger and PDL >1 machines in a data center warsteiner.db.cs.cmu.edu/demo/intemon.jsp SDM'7 Faloutsos, Kolda, Sun 1-16 P3: Social network analysis Traditionally, people focus on static networks and find community structures We plan to monitor the change of the community structure over time 199 Authors 24 DM Keywords DB SDM'7 Faloutsos, Kolda, Sun 1-17 DB P4: Network forensics Directional network flows A large ISP with 1 POPs, each POP 1Gbps link capacity [Hotnets24] 45 GB/hour with compression Task: Identify abnormal traffic pattern and find out the cause abnormal traffic normal traffic 5 45 destination source source destination source SDM'7 Faloutsos, Kolda, Sun Collaboration with Prof. Hui Zhang and Dr. Yinglian Xie 1-18 destination source 3

4 Faloutsos, Kolda, Sun SDM'7 P5: Web graph mining How to order the importance of web pages? Kleinberg s algorithm HITS PageRank Tensor extension on HITS (TOPHITS) context-sensitive hypergraph analysis Order Correspondence Static Data model Tensor Formally, Generalization of matrices Represented as multi-array, (~ data cube). 1st Vector 2 nd Matrix 3 rd 3D array Example Sensors Authors Keywords Destinations Ports SDM'7 Faloutsos, Kolda, Sun 1-19 Sources SDM'7 Faloutsos, Kolda, Sun 1-2 Order Correspondence Example Dynamic Data model Tensor Streams A sequence of Mth order tensor where t is increasing over time 1st Multiple streams Sensors time author Time evolving graphs 3 rd 3D arrays SDM'7 Faloutsos, Kolda, Sun nd keyword time Destinations Ports Sources Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap SDM'7 Faloutsos, Kolda, Sun

5 Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap SVD, PCA HITS, PageRank CUR Co-clustering Nonnegative Matrix factorization Patterns Anomaly detection Compression General goals SDM'7 Faloutsos, Kolda, Sun 2-1 SDM'7 Faloutsos, Kolda, Sun 2-2 Examples of Matrices Example/Intuition: Documents and terms Find patterns, groups, concepts Paper#1 Paper#2 Paper#3 Paper#4... data mining classif. tree SDM'7 Faloutsos, Kolda, Sun 2-3 Singular Value Decomposition (SVD) X x (1) x (2) x (M) input data U u 1 u 2 u k left singular vectors X UΣV T. σ 1 σ 2 SDM'7 Faloutsos, Kolda, Sun 2-4 Σ σ k singular values. v 1 v 2 v k V T right singular vectors SVD as spectral decomposition SVD - Example m n A n σ 1 u 1 v 1 σ 2 u 2 v 2 Σ m + V T U Best rank-k approximation in L2 and Frobenius SVD only works for static matrices (a single 2 nd order tensor) CS MD A U Σ V T - example: lung data infṛetrieval brain x x SDM'7 Faloutsos, Kolda, Sun 2-5 See also PARAFAC SDM'7 Faloutsos, Kolda, Sun 2-6 1

6 SVD - Example A U Σ V T - example: data infṛetrieval CS-concept brain lung MD-concept CS x x MD SDM'7 Faloutsos, Kolda, Sun 2-7 SVD - Example A U Σ V T - example: doc-to-concept similarity matrix data infṛetrieval CS-concept brain lung MD-concept CS x x MD SDM'7 Faloutsos, Kolda, Sun 2-8 SVD - Example A U Σ V T - example: data infṛetrieval brain lung strength of CS-concept CS x x MD SDM'7 Faloutsos, Kolda, Sun 2-9 CS MD SVD - Example A U Σ V T - example: lung data infṛetrieval brain CS-concept SDM'7 Faloutsos, Kolda, Sun 2-1 x term-to-concept similarity matrix x CS MD SVD - Example A U Σ V T - example: lung data infṛetrieval brain CS-concept SDM'7 Faloutsos, Kolda, Sun 2-11 x term-to-concept similarity matrix x SVD properties V are the eigenvectors of the covariance matrix X T X, since U are the eigenvectors of the Gram (innerproduct) matrix XX T, since Further reading: 1. Ian T. Jolliffe, Principal Component Analysis (2 nd ed), Springer, 22. SDM'7 Faloutsos, Kolda, Sun 2. Gilbert Strang, Linear Algebra and Its Applications (4 th ed), Brooks Cole,

7 SVD - Interpretation documents, terms and concepts : Q: if A is the document-to-term matrix, what is A T A? A: term-to-term ([m x m]) similarity matrix Q: A A T? A: document-to-document ([n x n]) similarity matrix SDM'7 Faloutsos, Kolda, Sun 2-13 Principal Component Analysis (PCA) SVD m n A m k R U kr PCs PCA is an important application of SVD Note that U and V are dense and may have negative entries SDM'7 Faloutsos, Kolda, Sun 2-14 Σ n V T Loading kr PCA interpretation best axis to project on: ( best min sum of squares of projection errors) Term2 ( lung ) PCA - interpretation Term2 ( lung ) PCA projects points Onto the best axis first singular vector v1 minimum RMS error Term1 ( data ) Term1 ( data ) SDM'7 Faloutsos, Kolda, Sun 2-15 SDM'7 Faloutsos, Kolda, Sun 2-16 Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap SVD, PCA HITS, PageRank CUR Co-clustering Nonnegative Matrix factorization Kleinberg s algorithm HITS Problem dfn: given the web and a query find the most authoritative web pages for this query Step : find all pages containing the query terms Step 1: expand by one move forward and backward SDM'7 Faloutsos, Kolda, Sun 2-17 Further SDM'7 reading: Faloutsos, Kolda, Sun J. Kleinberg. Authoritative sources in a hyperlinked environment. SODA

8 Kleinberg s algorithm HITS Step 1: expand by one move forward and backward Kleinberg s algorithm HITS on the resulting graph, give high score ( authorities ) to nodes that many important nodes point to give high importance score ( hubs ) to nodes that point to good authorities hubs authorities SDM'7 Faloutsos, Kolda, Sun 2-19 SDM'7 Faloutsos, Kolda, Sun 2-2 Kleinberg s algorithm HITS observations recursive definition! each node (say, i -th node) has both an authoritativeness score a i and a hubness score h i Kleinberg s algorithm: HITS Let A be the adjacency matrix: the (i,j) entry is 1 if the edge from i to j exists Let h and a be [n x 1] vectors with the hubness and authoritativiness scores. Then: SDM'7 Faloutsos, Kolda, Sun 2-21 SDM'7 Faloutsos, Kolda, Sun 2-22 Kleinberg s algorithm: HITS Kleinberg s algorithm: HITS k l m i Then: a i h k + h l + h m that is a i Sum (h j ) over all j that (j,i) edge exists or a A T h i n q p symmetrically, for the hubness : h i a n + a p + a q that is h i Sum (q j ) over all j that (i,j) edge exists or h Aa SDM'7 Faloutsos, Kolda, Sun 2-23 SDM'7 Faloutsos, Kolda, Sun

9 Kleinberg s algorithm: HITS In conclusion, we want vectors h and a such that: h A a a A T h That is: a A T A a SDM'7 Faloutsos, Kolda, Sun 2-25 Kleinberg s algorithm: HITS a is a right singular vector of the adjacency matrix A (by dfn!), a.k.a the eigenvector of A T A Starting from random a and iterating, we ll eventually converge Q: to which of all the eigenvectors? why? A: to the one of the strongest eigenvalue, (A T A ) k v ~ (constant) v 1 SDM'7 Faloutsos, Kolda, Sun 2-26 Kleinberg s algorithm - discussion authority score can be used to find similar pages (how?) closely related to citation analysis, social networks / small world phenomena Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap SVD, PCA HITS,PageRank CUR Co-clustering Nonnegative Matrix factorization See also TOPHITS SDM'7 Faloutsos, Kolda, Sun 2-27 SDM'7 Faloutsos, Kolda, Sun 2-28 Motivating problem: PageRank Given a directed graph, find its most interesting/central node A node is important, if it is connected with important nodes (recursive, but OK!) SDM'7 Faloutsos, Kolda, Sun 2-29 Motivating problem PageRank solution Given a directed graph, find its most interesting/central node Proposed solution: Random walk; spot most popular node (-> steady state prob. (ssp)) A node has high ssp, if it is connected with high ssp nodes (recursive, but OK!) SDM'7 Faloutsos, Kolda, Sun 2-3 5

10 (Simplified) PageRank algorithm Let A be the transition matrix ( adjacency matrix); let A T become column-normalized - then From A T To 1 p p2 3 1/2 1/2 p3 4 5 p1 p2 p3 1/2 p4 p4 1/2 p5 p5 (Simplified) PageRank algorithm A T p p A T p p p1 1 1 p2 1/2 1/2 1/2 1/2 p3 p4 p5 p1 p2 p3 p4 p5 SDM'7 Faloutsos, Kolda, Sun 2-31 SDM'7 Faloutsos, Kolda, Sun 2-32 (Simplified) PageRank algorithm A T p 1 * p thus, p is the eigenvector that corresponds to the highest eigenvalue (1, since the matrix is column-normalized) Why does it exist such a p? p exists if A is nxn, nonnegative, irreducible [Perron Frobenius theorem] (Simplified) PageRank algorithm In short: imagine a particle randomly moving along the edges compute its steady-state probabilities (ssp) Full version of algo: with occasional random jumps Why? To make the matrix irreducible SDM'7 Faloutsos, Kolda, Sun 2-33 SDM'7 Faloutsos, Kolda, Sun 2-34 Full Algorithm With probability 1-c, fly-out to a random node Then, we have p c A p+ (1-c)/n 1 > p (1-c)/n [I - c A] -1 1 Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap SVD, PCA HITS, PageRank CUR Co-clustering Nonnegative Matrix factorization SDM'7 Faloutsos, Kolda, Sun 2-35 SDM'7 Faloutsos, Kolda, Sun

11 Motivation of CUR or CMD SVD, PCA all transform data into some abstract space (specified by a set basis) Interpretability problem Loss of sparsity Interpretability problem Each column of projection matrix U i is a linear combination of all dimensions along certain mode U i (:,1) [.5; -.5;.5;.5] All the data are projected onto the span of U i It is hard to interpret the projections SDM'7 Faloutsos, Kolda, Sun 2-37 SDM'7 Faloutsos, Kolda, Sun 2-38 PCA projects points Onto the best axis minimum RMS error PCA - interpretation Term2 ( lung ) v1 first singular vector Term1 ( data ) SDM'7 Faloutsos, Kolda, Sun 2-39 The sparsity property sparse and small SVD: A U Σ V T Big but sparse Big and dense dense but small CUR: A C U R Big but sparse Big but sparse SDM'7 Faloutsos, Kolda, Sun 2-4 The sparsity property pictorially: The sparsity property (cont.) SVD/PCA: Destroys sparsity space ratio SVD CUR CMD space ratio 1 2 SVD CUR CMD U Σ V T C U R CUR: maintains sparsity SDM'7 Faloutsos, Kolda, Sun accuracy Network DBLP CMD uses much smaller space to achieve the same accuracy CUR limitation: duplicate columns and rows SVD limitation: orthogonal projection densifies the data Reference: SDM'7 Faloutsos, Kolda, Sun 2-42 Sun et al. Less is More: Compact Matrix Decomposition for Large Sparse Graphs, SDM accuracy 7

12 CUR Example-based projection: use actual rows and columns to specify the subspace Given a matrix A R m n, find three matrices C R m c, U R c r, R R r n, such that A-CUR is small m n A m X C n R c U is the pseudo-inverse of X SDM'7 Faloutsos, Kolda, Sun 2-43 r Example-based Orthogonal projection CUR (cont.) Key question: How to select/sample the columns and rows? Uniform sampling Biased sampling CUR w/ absolute error bound CUR w/ relative error bound Reference: 1. Tutorial: Randomized Algorithms for Matrices and Massive Datasets, SDM 6 2. Drineas et al. Subspace Sampling and Relative-error Matrix Approximation: Column- Row-Based Methods, ESA26 3. Drineas SDM'7et al., Fast Monte Carlo Algorithms Faloutsos, Kolda, for Sun Matrices III: Computing a 2-44 Compressed Approximate Matrix Decomposition, SIAM Journal on Computing, 26. Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap SVD, PCA HITS, PageRank CUR Co-clustering etc Nonnegative Matrix factorization Co-clustering Let X and Y be discrete random variables X and Y take values in {1, 2,, m} and {1, 2,, n} p(x, Y) denotes the joint probability distribution if not known, it is often estimated based on co-occurrence data Application areas: text mining, market-basket analysis, analysis of browsing behavior, etc. Key Obstacles in Clustering Contingency Tables High Dimensionality, Sparsity, Noise Need for robust and scalable algorithms SDM'7 Faloutsos, Kolda, Sun 2-45 Reference: SDM'7 Faloutsos, Kolda, Sun Dhillon et al. Information-Theoretic Co-clustering, KDD 3 Co-clustering Given data matrix and the number of row and column groups k and l Simultaneously Cluster rows of p(x, Y) into k disjoint groups Cluster columns of p(x, Y) into l disjoint groups Key goal is to exploit the duality between row and column clustering to overcome sparsity and noise SDM'7 Faloutsos, Kolda, Sun 2-47 Information Theory Concepts Entropy of a random variable X with probability distribution p: H ( p) p( x) log p( x) x The Kullback-Leibler (KL) Divergence or Relative Entropy between two probability distributions p and q: KL ( p, q) p( x)log( p( x) q( x)) x Mutual Information between random variables X and Y: I( X, Y ) p( x, y)log p( x, y) details x y p( x) p( y) SDM'7 Faloutsos, Kolda, Sun

13 Information-Theoretic Co-Clustering View (scaled) co-occurrence matrix as a joint probability distribution between row & column random variables # co occurence( x, y) p( x, y) # co occurence( x, y) Y x, y Yˆ X We seek a hard-clustering of both dimensions such that loss in Mutual Information I ( X, Y ) - I ( Xˆ, Yˆ ) is minimized given a fixed no. of row & col. clusters SDM'7 Faloutsos, Kolda, Sun 2-49 Xˆ details Information Theoretic Co-clustering Loss in mutual information equals details I ( X, Y ) - I ( Xˆ, Yˆ) KL( p( x, y) q( x, y)) H ( Xˆ, Yˆ) + H ( X X ˆ ) + H ( Y Yˆ) H ( X, Y ) p is the input distribution q is an approximation to p q( x, y) p(ˆ, x yˆ) p( x xˆ) p( y yˆ), x xˆ, y yˆ Can be shown that q(x,y) is a maximum entropy approximation subject to cluster constraints. SDM'7 Faloutsos, Kolda, Sun p( x, y) p( x xˆ) [ ].2.2 p( xˆ, yˆ) p( y yˆ) SDM'7 Faloutsos, Kolda, Sun q( x, y) #parameters that determine q(x,y) are: ( m k) + ( kl 1) + ( n l) Problem with Information Theoretic Co-clustering Number of row and column groups must be specified Desiderata: Simultaneously discover row and column groups Fully Automatic: No magic numbers Scalable to large graphs SDM'7 Faloutsos, Kolda, Sun 2-52 Cross-association What makes a cross-association good? Row groups versus Row groups Why is this better? Desiderata: Simultaneously discover row and column groups Fully Automatic: No magic numbers Scalable to large matrices Column groups Column groups Reference: SDM'7 Faloutsos, Kolda, Sun Chakrabarti et al. Fully Automatic Cross-Associations, KDD 4 SDM'7 Faloutsos, Kolda, Sun

14 What makes a cross-association good? What makes a cross-association good? Row groups Column groups versus Row groups Column groups Why is this better? simpler; easier to describe easier to compress! Problem definition: given an encoding scheme decide on the # of col. and row groups k and l and reorder rows and columns, to achieve best compression SDM'7 Faloutsos, Kolda, Sun 2-55 SDM'7 Faloutsos, Kolda, Sun 2-56 details Main Idea Algorithm Good Compression Total Encoding Cost Σ size i * H(x i ) + i Code Cost Better Clustering Cost of describing cross-associations Description Cost l 5 col groups k 5 row groups Minimize the total cost (# bits) for lossless compression k1, l2 k2, l2 k2, l3 k3, l3 k3, l4 k4, l4 k4, l5 SDM'7 Faloutsos, Kolda, Sun 2-57 SDM'7 Faloutsos, Kolda, Sun 2-58 Algorithm Code for cross-associations (matlab): Cross-Associations vs. Co-clustering Information-theoretic co-clustering Cross-Associations ations tgz Variations and extensions: Autopart [Chakrabarti, PKDD 4] 1. For any nonnegative matrix 2. Lossy Compression. 3. Approximates the original matrix, while trying to minimize KLdivergence. 4. The number of row and column groups must be given by the user. 1. For binary matrix 2. Lossless Compression. 3. Always provides complete information about the matrix, for any number of row and column groups. 4. Chosen automatically using the MDL principle. SDM'7 Faloutsos, Kolda, Sun 2-59 SDM'7 Faloutsos, Kolda, Sun 2-6 1

15 Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap SVD, PCA HITS, PageRank CUR Co-clustering, etc Nonnegative Matrix factorization Nonnegative Matrix Factorization Coming up soon with nonnegative tensor factorization SDM'7 Faloutsos, Kolda, Sun 2-61 SDM'7 Faloutsos, Kolda, Sun

16 Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap Tensor Basics Tucker Tucker 1 (PCA) Tucker 2 Tucker 3 (HOSVD) PARAFAC Tensor Basics 3-1 A tensor is a multidimensional array An I J K tensor K Column (Mode-1) Fibers Row (Mode-2) Fibers Tube (Mode-3) Fibers Matricize (unfolding) Matricize: Converting a Tensor to a Matrix (i,j,k) (i,j ) X (n) : The mode-n fibers are rearranged to be the columns of a matrix I x ijk Reverse Matricize (i,j ) (i,j,k) J 3 rd order tensor mode 1 has dimension I mode 2 has dimension J mode 3 has dimension K Note: Tutorial focus is on 3 dimensions, but everything can be extended to higher dimensionality. Horizontal Slices Lateral Slices Frontal Slices Tensor Mode-n Multiplication Pictorial View of Mode-n Matrix Multiplication Tensor Times Matrix Tensor Times Vector Mode-2 multiplication (lateral slices) Multiply each row (mode-2) fiber by B Compute the dot product of a and each column (mode-1) fiber 3-5 Mode-1 multiplication (frontal slices) Mode-3 multiplication (horizontal slices) 3-6 1

17 Mode-n product Example Tensor times a matrix Mode-n product Example Tensor times a vector Location Type Time time Time clusters Location Type clusters Location Type Time time Time Location Type Outer Product Outer, Kronecker, & Khatri-Rao Products Kronecker Product M x N P x Q Specially Structured Tensors MP x NQ Khatri-Rao Product M x R N x R MN x R Observe: a b and a b have the same elements, but one is shaped into a matrix and the other into a vector. 3-9 Specially Structured Tensors Tucker Tensor Kruskal Tensor Specially Structured Tensors Tucker Tensor Kruskal Tensor core I x J x K I x R W K x T J x S I x J x K I x R K x R W J x R In matrix form: In matrix form: U V R x S x T U + + V R x R x R

18 Matrix SVD is a Tucker or Kruskal Tensor Matrix SVD: Tensor Decompositions Tucker Tensor: Kruskal Tensor: 3-13 I x J x K Tucker Decomposition I x R J x S A K x T C B R x S x T Tucker Variations See Kroonenberg & De Leeuw, Psychometrika, 198 for discussion. Tucker2 I x J x K I x R J x S B A R x S x K Proposed by Tucker (1966) AKA: Three-mode factor analysis, three-mode PCA, orthogonal array decomposition A, B, and C generally assumed to be orthonormal (generally assume they have full column rank) is not diagonal Not unique Recall the equations for converting a tensor to a matrix 3-15 Tucker1 I x J x K I x R A R x J x K Finding principal components in only mode 1. Can be solved via rank-r matrix SVD 3-16 I x J x K I x R A C K x T R x S x T Higher Order SVD (HO-SVD) J x S B Not optimal, but often used to initialize Tucker- ALS algorithm. (Observe connection to Tucker1.) Minimize s.t. A,B,C orthonormal Solving for Tucker Maximize s.t. A,B,C orthonormal To solve for A (assuming B and C are fixed): De Lathauwer, De Moor, & Vandewalle, SIMAX, Calculate R leading left singular vectors of

19 Alternating Least Squares (ALS) for Tucker I x J x K I x R A C K x T R x S x T J x S B Initialize Choose R, S, T Calculate A, B, C via HO-SVD Until converged do A R leading left singular vectors of X (1) (C B) B S leading left singular vectors of X (2) (C A) C T leading left singular vectors of X (3) (B A) Kroonenberg & De Leeuw, Psychometrika, CANDECOMP/PARAFAC Decomposition I x J x K I x R A C K x R R x R x R J x R CANDECOMP Canonical Decomposition (Carroll & Chang, 197) PARAFAC Parallel Factors (Harshman, 197) Core is diagonal (specified by the vector λ) Columns of A, B, and C are not orthonormal If R is minimal, then R is called the rank of the tensor (Kruskal 1977) Can have rank( ) > min{i,j,k} 3-2 B + + Alternating Least Squares (ALS) for PARAFAC + + To solve for A, consider: I x J x K R rank-1 tensors By the properties of the Khatri-Rao product, we have: Thus: Do the same for B and C, then repeat

20 Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap Other decompositions Nonnegativity Missing values Incrementalization Other Tensor Decompositions 4-1 M x N x P Combining Tucker & PARAFAC M x I N x J V Step 1: Choose orthonormal matrices U, V, W to compress tensor (Tucker tensor!) Typically HO-SVD can be U used I x J x K Step 2: Run PARAFAC on smaller tensor W P x K Step 3: Reassemble result 3-Way DEDICOM roles time patterns 3-way DEDICOM 2-way DEDICOM introduced by Harshman, way DEDICOM due to Kiers, 1993 Idea is to capture asymmetric relationships among different roles If third dimension is time, than diagonal slices capture participation of each role at each time Bro and Andersson, See, e.g., Bader, Harshman, Kolda, SAND Dense Tensors Computations with Tensors Largest tensor that can be stored on a laptop is 2 x 2 x 2 Typically, tensor operations are reduced to matrix operations Requires permuting and reshaping the tensor Example: Mode-n tensormatrix multiply I x J x K Example: Mode-1 Matrix Multiply M x J x K I x J x K M x I M x JK M x I I x JK 4-6 1

21 Sparse Tensors: Only Store Nonzeros Example: Tensor-Vector Multiply (in all modes) Tucker Tensors: Store Core & Factors Tucker tensor stores the core (which can be dense, sparse, or structured) and the factors. Store just the nonzeros of a tensor (assume coordinate format) pth nonzero 1 st subscript of pth nonzero 2 nd subscript of pth nonzero 3 rd subscript of pth nonzero 4-7 Example: Mode-3 Tensor-Vector Multiply Result is a Tucker Tensor 4-8 I x J x K Kruskal Example: Store Factors W K x R I x R J x R V U R x R x R Kruskal tensors store factor matrices and scaling vector. + + Nonnegativity Example: Norm I x R J x R K x R I x J x K R x R R x R R x R 4-9 Non-negative Matrix Factorization Non-negative 3-Way PARAFAC Factorization Update formulas (do not increase objective function): Minimize subject to elements of A and B being positive. Minimize subject to elements of A, B and C being positive. Lee-Seung-like update formulas can be derived for 3D and higher: Elementwise multiply (Hadamard product) Elementwise divide Elementwise multiply (Hadamard product) Elementwise divide Lee & Seung, Nature, M. Mørup, L. K. Hansen, J. Parnas, S. M. Arnfred, Decomposing the time-frequency representation of EEG using non-negative 4-12 matrix and multi-way factorization, 26 2

22 Handling Missing Data A Quick Overview on Handling Missing Data Consider sparse PARAFAC where all the zero entries represent missing data Typically, missing values are just set to zero There are more sophisticated approaches for handling missing values: Weighted approximation Data imputation to estimate missing values 4-14 See, e.g., Kiers, Psychometrika, 1997 and Srebro & Jaakkola, ICML 23 Weighted Least Squares Missing Value Imputation Weight Tensor Use the current estimate to fill in the missing values Current Estimate Weight the least squares problem so that the missing elements are ignored: The tensor for the next iteration of the algorithm is: Known Values Estimates of Unknowns Weighted Least Squares But this problem is often too hard to solve directly! 4-15 Sparse! Kruskal Tensor Challenge is finding a good initial estimate 4-16 Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap Tensor Toolbox for MATLAB Six object-oriented tensor classes Working with tensors is easy Most comprehensive set of kernel operations in any language E.g., arithmetic, logical, multiplication operations Sparse tensors are unique Speed-ups of two orders of magnitude for smaller problems Larger problems than ever before Free for research or evaluations purposes 297 unique registered users from all over the world (as of January 17, 26) 4-17 Bader & Kolda, ACM TOMS 26 & SAND

23 Incrementalization Incremental Tensor Decomposition Dynamic data model Tensor Streams Dynamic Tensor Decomposition (DTA) Streaming Tensor Decomposition (STA) Window-based Tensor Decomposition (WTA) 4-2 Destination Source Dynamic Tensor Stream time Streams come with structure (time, source, destination, port) (time, author, keyword) How to summarize tensor streams effectively and incrementally? 4-21 Order Correspondence Example Dynamic Data model Tensor Streams A sequence of Mth order tensor where n is increasing over time 1st Multiple streams Sensors time author 2 nd Time evolving graphs keyword time Destinations Ports 3 rd 3D arrays Sources 4-22 Incremental Tensor Decomposition Destination Source Old Tensors New Tensor Old cores U Destination 1 st order DTA - problem Given x 1 x n where each x i R N, find U R N R such that the error e is small: time n N x 1.? Y U T Sensors R x n Sensors indoor outdoor Note that Y XU U Source

24 T time 1 st order Dynamic Tensor Analysis Input: new data vector x R N, old variance matrix C R N N Output: new projection matrix U R N R Old X x Algorithm: 1. update variance matrix C new x T x+ C 2. Diagonalize UΛU T C new 3. Determine the rank R and return U M th order DTA Reconstruct Variance Matrix T U d U d S d C d Construct Variance Matrix of Incremental Tensor Matricizing Diagonalize Variance Matrix x x T C U C new U T Matricizing, Transpose X(d)X(d) C d Update Variance Matrix T U d U d S d Diagonalization has to be done for every new x! T X( d ) X( d ) M th order DTA complexity Storage: O( N i ), i.e., size of an input tensor at a single timestamp Computation: N i3 (or N i2 ) diagonalization of C + N i N i matrix multiplication X (d)t X (d) For low order tensor(<3), diagonalization is the main cost For high order tensor, matrix multiplication is the main cost st order Streaming Tensor Analysis (STA) Adjust U smoothly when new data arrive without diagonalization [VLDB5] For each new point x Project onto current line Estimate error Rotate line in the direction of the error and in proportion to its magnitude For each new point x and for i 1,, k : y i : U it x (proj. onto U i ) d i λd i + y 2 i (energy i-th eigenval.) e i : x y i U i (error) U i U i + (1/d i ) y i e i (update estimate) x x y i U i (repeat with remainder) U error Sensor Sensor 2 T X( d ) Matricizing M th order STA Meta-algorithm for window-based tensor analysis Type Run 1 st order STA along each mode x U Complexity: 1 updated e Storage: O( N i ) 1 U 1 Computation: R i N i which is smaller than DTA y 1 R 1R2 R Location N 2 Y Time U Type W D U Time N U Location 4-3 5

25 Moving Window scheme (MW) Update the variance matrix C (i) incrementally Diagonalize C(i) to find U(i) old C d Tensor Streams D n-w... D(n-1,W) D(n,W) Time D n new C d... A good and efficient initialization Update variance matrix U (d) Diagonalize

26 Motivation Matrix tools Tensor basics Tensor extensions Software demo Case studies Roadmap value value P1: Environmental sensor monitoring time (min) Temperature value value time (min) 2 1 Light 1.5 SDM'7 Faloutsos, Kolda, Sun time (min) time (min) Humidity Voltage SDM'7 Faloutsos, Kolda, Sun 5-2 Location Type 1 st factor Scaling factor 25 Time value P2: sensor monitoring time time (min) value location Volt Humid Temp Light type SDM'7 Faloutsos, Kolda, Sun location value type 1 st factor consists of the main trends: Daily periodicity on time Uniform on all locations Temp, Light and Volt are positively correlated while negatively correlated with Humid 2 nd factor Scaling factor 154 P2: sensor monitoring value time time (min) location Volt Humid Temp Light type 2 nd factor captures an atypical trend: Uniformly across all time Concentrating on 3 locations Mainly due to voltage Interpretation: two sensors have low battery, and the other one has high battery. SDM'7 Faloutsos, Kolda, Sun 5-4 value type P3: Social network analysis Multiway latent semantic indexing (LSI) Monitor the change of the community structure over time time 199 Authors 24 DB DM Keywords DB Pattern UK UA 199 Query Philip Yu Michael Stonebreaker 24 P3: Social network analysis (cont.) Authors michael carey, michael stonebreaker, h. jagadish, hector garcia-molina surajit chaudhuri,mitch cherniack,michael stonebreaker,ugur etintemel jiawei han,jian pei,philip s. yu, jianyong wang,charu c. aggarwal DB Keywords queri,parallel,optimization,concurr, objectorient distribut,systems,view,storage,servic,process, cache streams,pattern,support, cluster, index,gener,queri DM Year Two groups are correctly identified: Databases and Data mining People and concepts are drifting over time SDM'7 Faloutsos, Kolda, Sun 5-5 SDM'7 Faloutsos, Kolda, Sun 5-6 1

27 destination P4: Network anomaly detection source error hours source Abnormal traffic Reconstruction error Normal traffic over time Reconstruction error gives indication of anomalies. Prominent difference between normal and abnormal ones is mainly due to the unusual scanning activity (confirmed by the campus admin). destination P5: Web graph mining How to order the importance of web pages? Kleinberg s algorithm HITS PageRank Tensor extension on HITS (TOPHITS) SDM'7 Faloutsos, Kolda, Sun 5-7 SDM'7 Faloutsos, Kolda, Sun 5-8 Kleinberg s Hubs and Authorities (the HITS method) Sparse adjacency matrix and its SVD: authority scores for 1 st topic hub scores for 1 st topic authority scores for 2 nd topic hub scores for 2 nd topic SDM'7 Faloutsos, Kolda, Sun 5-9 Kleinberg, JACM, 1999 from to from HITS Authorities on Sample Data 1st Principal Factor nd Principal Factor.8 www-128.ibm.com.99 We started our crawl from.5 www2.lehigh.edu 3rd Principal Factor.6 and crawled 47 pages,.1 java.sun.com.6 resulting in 56.1 news.com.com cross-linked hosts..36 developers.sun.com.2 4th Principal Factor.24 see.sun.com.2 lewisweb.cc.lehigh.edu docs.sun.com blogs.sun.com.2 fp1.cc.lehigh.edu.31 travel.state.gov 6th Principal Factor.8 sunsolve.sun.com.22 mathpost.asu.edu math.la.asu.edu.8 news.com.com authority scores authority scores for 2 nd topic for 1 st topic.13 archives.math.utk.edu.2 to hub scores SDM'7 hub scores for 1 st topic for 2 nd topic Faloutsos, Kolda, Sun 5-1 Three-Dimensional View of the Web Topical HITS (TOPHITS) Main Idea: Extend the idea behind the HITS model to incorporate term (i.e., topical) information. Observe that this tensor is very sparse! term term scores for 1 st topic term scores for 2 nd topic to SDM'7 Faloutsos, Kolda, Sun 5-11 Kolda, Bader, Kenny, ICDM5 from hub scores for 1 st topic authority scores for 1 st topic authority scores for 2 nd topic hub scores for 2 nd topic SDM'7 Faloutsos, Kolda, Sun

28 Topical HITS (TOPHITS) Main Idea: Extend the idea behind the HITS model to incorporate term (i.e., topical) information. from term to hub scores for 1 st topic term scores for 1 st topic authority scores for 1 st topic term scores for 2 nd topic authority scores for 2 nd topic hub scores for 2 nd topic SDM'7 Faloutsos, Kolda, Sun 5-13 TOPHITS Terms & Authorities on Sample Data term scores term scores to authority scores for 2 nd topic hub scores for 2 scores topic 1 st topic 1st Principal Factor.86 java.sun.com.23 JAVA.18 SUN.38 developers.sun.com 2nd Principal Factor TOPHITS uses 3D analysis to find.17 PLATFORM.16 docs.sun.com.2 NO-READABLE-TEXT.99 SOLARIS.14 see.sun.com.16 the dominant groupings of web.16 FACULTY.6 www2.lehigh.edu 3rd Principal Factor.16 DEVELOPER.14 pages and terms..16 SEARCH.3 EDITION.15 NO-READABLE-TEXT NEWS DOWNLOAD.15 IBM.7 developer.sun.com LIBRARIES 4th Principal Factor www-128.ibm.com.14 INFO SERVICES.6 sunsolve.sun.com COMPUTING.12 WEBSPHERE access1.sun.com.26 INFORMATION SOFTWARE.5.12 LEHIGH iforce.sun.com.24 FEDERAL NO-READABLE-TEXT.12 WEB DEVELOPERWORKS.1 6th Principal.23 CITIZEN Factor w k # unique links using term k.11 LINUX.22 OTHER.19 travel.state.gov.26 PRESIDENT.87 RESOURCES.19 CENTER.18 NO-READABLE-TEXT.18 TECHNOLOGIES.19 LANGUAGES.9 BUSH.16 travel.state.gov 12th Principal Factor.1 DOWNLOADS.15 U.S WELCOME PUBLICATIONS.75 OPTIMIZATION WHITE.8 CONSUMER SOFTWARE U.S.5 FREE DECISION th plato.la.asu.edu Principal Factor.4 HOUSE.7 NEOS ADOBE BUDGET.45 READER.6 TREE.28 PRESIDENTS.4 ACROBAT.5 GUIDE th Principal Factor.11 OFFICE.2 SEARCH.3 FREE.5 WEATHER Tensor PARAFAC ENGINE.3 NO-READABLE-TEXT.24 OFFICE.25 www-fp.mcs.anl.gov CONTROL.29 HERE.5 CENTER.22 lwf.ncdc.noaa.gov 19th Principal Factor.29 COPY.5 ILOG NO-READABLE-TEXT for 2 nd topic ORGANIZATION.22 TAX DOWNLOAD.17 for 1 st topic TAXES travel.state.gov.15 NWS.9.15 SEVERE.7 aviationweather.gov.22 CHILD.15 RETIREMENT FIRE.15 POLICY BENEFITS authority scores for 1 st topic CLIMATE.14 STATE.3 INCOME.3 SDM'7hub for Faloutsos, Kolda, Sun SERVICE.2.13 REVENUE.2 CREDIT.1 from term Conclusion Real data are often in high dimensions with multiple aspects (modes) Matrix and tensor provide elegant theory and algorithms for such data However, many problems are still open skew distribution, anomaly detection, streaming algorithm, distributed/parallel algorithms, efficient out-of-core processing Thank you! Christos Faloutsos Tamara Kolda csmr.ca.sandia.gov/~tgkolda Jimeng Sun SDM'7 Faloutsos, Kolda, Sun 5-15 SDM'7 Faloutsos, Kolda, Sun

Large Graph Mining: Power Tools and a Practitioner s guide

Large Graph Mining: Power Tools and a Practitioner s guide Large Graph Mining: Power Tools and a Practitioner s guide Task 5: Graphs over time & tensors Faloutsos, Miller, Tsourakakis CMU KDD '9 Faloutsos, Miller, Tsourakakis P5-1 Outline Introduction Motivation

More information

Faloutsos, Tong ICDE, 2009

Faloutsos, Tong ICDE, 2009 Large Graph Mining: Patterns, Tools and Case Studies Christos Faloutsos Hanghang Tong CMU Copyright: Faloutsos, Tong (29) 2-1 Outline Part 1: Patterns Part 2: Matrix and Tensor Tools Part 3: Proximity

More information

Must-read Material : Multimedia Databases and Data Mining. Indexing - Detailed outline. Outline. Faloutsos

Must-read Material : Multimedia Databases and Data Mining. Indexing - Detailed outline. Outline. Faloutsos Must-read Material 15-826: Multimedia Databases and Data Mining Tamara G. Kolda and Brett W. Bader. Tensor decompositions and applications. Technical Report SAND2007-6702, Sandia National Laboratories,

More information

Higher-Order Web Link Analysis Using Multilinear Algebra

Higher-Order Web Link Analysis Using Multilinear Algebra Higher-Order Web Link Analysis Using Multilinear Algebra Tamara G. Kolda, Brett W. Bader, and Joseph P. Kenny Sandia National Laboratories Livermore, CA and Albuquerque, NM {tgkolda,bwbader,jpkenny}@sandia.gov

More information

Window-based Tensor Analysis on High-dimensional and Multi-aspect Streams

Window-based Tensor Analysis on High-dimensional and Multi-aspect Streams Window-based Tensor Analysis on High-dimensional and Multi-aspect Streams Jimeng Sun Spiros Papadimitriou Philip S. Yu Carnegie Mellon University Pittsburgh, PA, USA IBM T.J. Watson Research Center Hawthorne,

More information

Tensor Analysis. Topics in Data Mining Fall Bruno Ribeiro

Tensor Analysis. Topics in Data Mining Fall Bruno Ribeiro Tensor Analysis Topics in Data Mining Fall 2015 Bruno Ribeiro Tensor Basics But First 2 Mining Matrices 3 Singular Value Decomposition (SVD) } X(i,j) = value of user i for property j i 2 j 5 X(Alice, cholesterol)

More information

Incremental Pattern Discovery on Streams, Graphs and Tensors

Incremental Pattern Discovery on Streams, Graphs and Tensors Incremental Pattern Discovery on Streams, Graphs and Tensors Jimeng Sun Thesis Committee: Christos Faloutsos Tom Mitchell David Steier, External member Philip S. Yu, External member Hui Zhang Abstract

More information

CS60021: Scalable Data Mining. Dimensionality Reduction

CS60021: Scalable Data Mining. Dimensionality Reduction J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org 1 CS60021: Scalable Data Mining Dimensionality Reduction Sourangshu Bhattacharya Assumption: Data lies on or near a

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 6: Some Other Stuff PD Dr.

More information

Scalable Tensor Factorizations with Incomplete Data

Scalable Tensor Factorizations with Incomplete Data Scalable Tensor Factorizations with Incomplete Data Tamara G. Kolda & Daniel M. Dunlavy Sandia National Labs Evrim Acar Information Technologies Institute, TUBITAK-UEKAE, Turkey Morten Mørup Technical

More information

Parallel Tensor Compression for Large-Scale Scientific Data

Parallel Tensor Compression for Large-Scale Scientific Data Parallel Tensor Compression for Large-Scale Scientific Data Woody Austin, Grey Ballard, Tamara G. Kolda April 14, 2016 SIAM Conference on Parallel Processing for Scientific Computing MS 44/52: Parallel

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Lecture 6: Numerical Linear Algebra: Applications in Machine Learning Cho-Jui Hsieh UC Davis April 27, 2017 Principal Component Analysis Principal

More information

Introduction to Tensors. 8 May 2014

Introduction to Tensors. 8 May 2014 Introduction to Tensors 8 May 2014 Introduction to Tensors What is a tensor? Basic Operations CP Decompositions and Tensor Rank Matricization and Computing the CP Dear Tullio,! I admire the elegance of

More information

Two heads better than one: Pattern Discovery in Time-evolving Multi-Aspect Data

Two heads better than one: Pattern Discovery in Time-evolving Multi-Aspect Data Two heads better than one: Pattern Discovery in Time-evolving Multi-Aspect Data Jimeng Sun 1, Charalampos E. Tsourakakis 2, Evan Hoke 4, Christos Faloutsos 2, and Tina Eliassi-Rad 3 1 IBM T.J. Watson Research

More information

CS 3750 Advanced Machine Learning. Applications of SVD and PCA (LSA and Link analysis) Cem Akkaya

CS 3750 Advanced Machine Learning. Applications of SVD and PCA (LSA and Link analysis) Cem Akkaya CS 375 Advanced Machine Learning Applications of SVD and PCA (LSA and Link analysis) Cem Akkaya Outline SVD and LSI Kleinberg s Algorithm PageRank Algorithm Vector Space Model Vector space model represents

More information

The multiple-vector tensor-vector product

The multiple-vector tensor-vector product I TD MTVP C KU Leuven August 29, 2013 In collaboration with: N Vanbaelen, K Meerbergen, and R Vandebril Overview I TD MTVP C 1 Introduction Inspiring example Notation 2 Tensor decompositions The CP decomposition

More information

Text Analytics (Text Mining)

Text Analytics (Text Mining) http://poloclub.gatech.edu/cse6242 CSE6242 / CX4242: Data & Visual Analytics Text Analytics (Text Mining) Concepts, Algorithms, LSI/SVD Duen Horng (Polo) Chau Assistant Professor Associate Director, MS

More information

Fitting a Tensor Decomposition is a Nonlinear Optimization Problem

Fitting a Tensor Decomposition is a Nonlinear Optimization Problem Fitting a Tensor Decomposition is a Nonlinear Optimization Problem Evrim Acar, Daniel M. Dunlavy, and Tamara G. Kolda* Sandia National Laboratories Sandia is a multiprogram laboratory operated by Sandia

More information

1 Searching the World Wide Web

1 Searching the World Wide Web Hubs and Authorities in a Hyperlinked Environment 1 Searching the World Wide Web Because diverse users each modify the link structure of the WWW within a relatively small scope by creating web-pages on

More information

Fundamentals of Multilinear Subspace Learning

Fundamentals of Multilinear Subspace Learning Chapter 3 Fundamentals of Multilinear Subspace Learning The previous chapter covered background materials on linear subspace learning. From this chapter on, we shall proceed to multiple dimensions with

More information

Lecture 12: Link Analysis for Web Retrieval

Lecture 12: Link Analysis for Web Retrieval Lecture 12: Link Analysis for Web Retrieval Trevor Cohn COMP90042, 2015, Semester 1 What we ll learn in this lecture The web as a graph Page-rank method for deriving the importance of pages Hubs and authorities

More information

Introduction to Data Mining

Introduction to Data Mining Introduction to Data Mining Lecture #21: Dimensionality Reduction Seoul National University 1 In This Lecture Understand the motivation and applications of dimensionality reduction Learn the definition

More information

Tutorial on MATLAB for tensors and the Tucker decomposition

Tutorial on MATLAB for tensors and the Tucker decomposition Tutorial on MATLAB for tensors and the Tucker decomposition Tamara G. Kolda and Brett W. Bader Workshop on Tensor Decomposition and Applications CIRM, Luminy, Marseille, France August 29, 2005 Sandia is

More information

The Canonical Tensor Decomposition and Its Applications to Social Network Analysis

The Canonical Tensor Decomposition and Its Applications to Social Network Analysis The Canonical Tensor Decomposition and Its Applications to Social Network Analysis Evrim Acar, Tamara G. Kolda and Daniel M. Dunlavy Sandia National Labs Sandia is a multiprogram laboratory operated by

More information

Motivation Data mining: ~ find patterns (rules, outliers) Problem#1: How do real graphs look like? Problem#2: How do they evolve? Problem#3: How to ge

Motivation Data mining: ~ find patterns (rules, outliers) Problem#1: How do real graphs look like? Problem#2: How do they evolve? Problem#3: How to ge Graph Mining: Laws, Generators and Tools Christos Faloutsos CMU Thank you! Sharad Mehrotra UCI 2007 C. Faloutsos 2 Outline Problem definition / Motivation Static & dynamic laws; generators Tools: CenterPiece

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu 2/7/2012 Jure Leskovec, Stanford C246: Mining Massive Datasets 2 Web pages are not equally important www.joe-schmoe.com

More information

Two heads better than one: pattern discovery in time-evolving multi-aspect data

Two heads better than one: pattern discovery in time-evolving multi-aspect data Data Min Knowl Disc (2008) 17:111 128 DOI 10.1007/s10618-008-0112-3 Two heads better than one: pattern discovery in time-evolving multi-aspect data Jimeng Sun Charalampos E. Tsourakakis Evan Hoke Christos

More information

Link Analysis. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze

Link Analysis. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze Link Analysis Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze 1 The Web as a Directed Graph Page A Anchor hyperlink Page B Assumption 1: A hyperlink between pages

More information

Information Retrieval

Information Retrieval Introduction to Information CS276: Information and Web Search Christopher Manning and Pandu Nayak Lecture 13: Latent Semantic Indexing Ch. 18 Today s topic Latent Semantic Indexing Term-document matrices

More information

Index. Copyright (c)2007 The Society for Industrial and Applied Mathematics From: Matrix Methods in Data Mining and Pattern Recgonition By: Lars Elden

Index. Copyright (c)2007 The Society for Industrial and Applied Mathematics From: Matrix Methods in Data Mining and Pattern Recgonition By: Lars Elden Index 1-norm, 15 matrix, 17 vector, 15 2-norm, 15, 59 matrix, 17 vector, 15 3-mode array, 91 absolute error, 15 adjacency matrix, 158 Aitken extrapolation, 157 algebra, multi-linear, 91 all-orthogonality,

More information

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia Network Feature s Decompositions for Machine Learning 1 1 Department of Computer Science University of British Columbia UBC Machine Learning Group, June 15 2016 1/30 Contact information Network Feature

More information

CS246 Final Exam, Winter 2011

CS246 Final Exam, Winter 2011 CS246 Final Exam, Winter 2011 1. Your name and student ID. Name:... Student ID:... 2. I agree to comply with Stanford Honor Code. Signature:... 3. There should be 17 numbered pages in this exam (including

More information

Tensor Decompositions and Applications

Tensor Decompositions and Applications Tamara G. Kolda and Brett W. Bader Part I September 22, 2015 What is tensor? A N-th order tensor is an element of the tensor product of N vector spaces, each of which has its own coordinate system. a =

More information

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS DATA MINING LECTURE 3 Link Analysis Ranking PageRank -- Random walks HITS How to organize the web First try: Manually curated Web Directories How to organize the web Second try: Web Search Information

More information

Numerical Methods I Singular Value Decomposition

Numerical Methods I Singular Value Decomposition Numerical Methods I Singular Value Decomposition Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 9th, 2014 A. Donev (Courant Institute)

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 12 Jan-Willem van de Meent (credit: Yijun Zhao, Percy Liang) DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Linear Dimensionality

More information

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa Introduction to Search Engine Technology Introduction to Link Structure Analysis Ronny Lempel Yahoo Labs, Haifa Outline Anchor-text indexing Mathematical Background Motivation for link structure analysis

More information

From Matrix to Tensor. Charles F. Van Loan

From Matrix to Tensor. Charles F. Van Loan From Matrix to Tensor Charles F. Van Loan Department of Computer Science January 28, 2016 From Matrix to Tensor From Tensor To Matrix 1 / 68 What is a Tensor? Instead of just A(i, j) it s A(i, j, k) or

More information

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors Nico Vervliet Joint work with Lieven De Lathauwer SIAM AN17, July 13, 2017 2 Classification of hazardous

More information

ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition

ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition Wing-Kin (Ken) Ma 2017 2018 Term 2 Department of Electronic Engineering The Chinese University

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Linear Algebra and Eigenproblems

Linear Algebra and Eigenproblems Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details

More information

Multi-Way Compressed Sensing for Big Tensor Data

Multi-Way Compressed Sensing for Big Tensor Data Multi-Way Compressed Sensing for Big Tensor Data Nikos Sidiropoulos Dept. ECE University of Minnesota MIIS, July 1, 2013 Nikos Sidiropoulos Dept. ECE University of Minnesota ()Multi-Way Compressed Sensing

More information

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Online Tensor Factorization for. Feature Selection in EEG

Online Tensor Factorization for. Feature Selection in EEG Online Tensor Factorization for Feature Selection in EEG Alric Althoff Honors Thesis, Department of Cognitive Science, University of California - San Diego Supervised by Dr. Virginia de Sa Abstract Tensor

More information

PROBABILISTIC LATENT SEMANTIC ANALYSIS

PROBABILISTIC LATENT SEMANTIC ANALYSIS PROBABILISTIC LATENT SEMANTIC ANALYSIS Lingjia Deng Revised from slides of Shuguang Wang Outline Review of previous notes PCA/SVD HITS Latent Semantic Analysis Probabilistic Latent Semantic Analysis Applications

More information

Data Mining and Matrices

Data Mining and Matrices Data Mining and Matrices 10 Graphs II Rainer Gemulla, Pauli Miettinen Jul 4, 2013 Link analysis The web as a directed graph Set of web pages with associated textual content Hyperlinks between webpages

More information

PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211

PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211 PV211: Introduction to Information Retrieval https://www.fi.muni.cz/~sojka/pv211 IIR 18: Latent Semantic Indexing Handout version Petr Sojka, Hinrich Schütze et al. Faculty of Informatics, Masaryk University,

More information

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data. Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two

More information

/16/$ IEEE 1728

/16/$ IEEE 1728 Extension of the Semi-Algebraic Framework for Approximate CP Decompositions via Simultaneous Matrix Diagonalization to the Efficient Calculation of Coupled CP Decompositions Kristina Naskovska and Martin

More information

Link Analysis Ranking

Link Analysis Ranking Link Analysis Ranking How do search engines decide how to rank your query results? Guess why Google ranks the query results the way it does How would you do it? Naïve ranking of query results Given query

More information

Introduction to Information Retrieval

Introduction to Information Retrieval Introduction to Information Retrieval http://informationretrieval.org IIR 18: Latent Semantic Indexing Hinrich Schütze Center for Information and Language Processing, University of Munich 2013-07-10 1/43

More information

CS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine

CS 277: Data Mining. Mining Web Link Structure. CS 277: Data Mining Lectures Analyzing Web Link Structure Padhraic Smyth, UC Irvine CS 277: Data Mining Mining Web Link Structure Class Presentations In-class, Tuesday and Thursday next week 2-person teams: 6 minutes, up to 6 slides, 3 minutes/slides each person 1-person teams 4 minutes,

More information

arxiv: v1 [cs.lg] 18 Nov 2018

arxiv: v1 [cs.lg] 18 Nov 2018 THE CORE CONSISTENCY OF A COMPRESSED TENSOR Georgios Tsitsikas, Evangelos E. Papalexakis Dept. of Computer Science and Engineering University of California, Riverside arxiv:1811.7428v1 [cs.lg] 18 Nov 18

More information

Thanks to Jure Leskovec, Stanford and Panayiotis Tsaparas, Univ. of Ioannina for slides

Thanks to Jure Leskovec, Stanford and Panayiotis Tsaparas, Univ. of Ioannina for slides Thanks to Jure Leskovec, Stanford and Panayiotis Tsaparas, Univ. of Ioannina for slides Web Search: How to Organize the Web? Ranking Nodes on Graphs Hubs and Authorities PageRank How to Solve PageRank

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Link Analysis Information Retrieval and Data Mining. Prof. Matteo Matteucci

Link Analysis Information Retrieval and Data Mining. Prof. Matteo Matteucci Link Analysis Information Retrieval and Data Mining Prof. Matteo Matteucci Hyperlinks for Indexing and Ranking 2 Page A Hyperlink Page B Intuitions The anchor text might describe the target page B Anchor

More information

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Mining of Massive Datasets Jure Leskovec, AnandRajaraman, Jeff Ullman Stanford University

Mining of Massive Datasets Jure Leskovec, AnandRajaraman, Jeff Ullman Stanford University Note to other teachers and users of these slides: We would be delighted if you found this our material useful in giving your own lectures. Feel free to use these slides verbatim, or to modify them to fit

More information

MultiRank and HAR for Ranking Multi-relational Data, Transition Probability Tensors, and Multi-Stochastic Tensors

MultiRank and HAR for Ranking Multi-relational Data, Transition Probability Tensors, and Multi-Stochastic Tensors MultiRank and HAR for Ranking Multi-relational Data, Transition Probability Tensors, and Multi-Stochastic Tensors Michael K. Ng Centre for Mathematical Imaging and Vision and Department of Mathematics

More information

CVPR A New Tensor Algebra - Tutorial. July 26, 2017

CVPR A New Tensor Algebra - Tutorial. July 26, 2017 CVPR 2017 A New Tensor Algebra - Tutorial Lior Horesh lhoresh@us.ibm.com Misha Kilmer misha.kilmer@tufts.edu July 26, 2017 Outline Motivation Background and notation New t-product and associated algebraic

More information

Thanks to Jure Leskovec, Stanford and Panayiotis Tsaparas, Univ. of Ioannina for slides

Thanks to Jure Leskovec, Stanford and Panayiotis Tsaparas, Univ. of Ioannina for slides Thanks to Jure Leskovec, Stanford and Panayiotis Tsaparas, Univ. of Ioannina for slides Web Search: How to Organize the Web? Ranking Nodes on Graphs Hubs and Authorities PageRank How to Solve PageRank

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

Slides based on those in:

Slides based on those in: Spyros Kontogiannis & Christos Zaroliagis Slides based on those in: http://www.mmds.org High dim. data Graph data Infinite data Machine learning Apps Locality sensitive hashing PageRank, SimRank Filtering

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu November 16, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision

More information

1. Ignoring case, extract all unique words from the entire set of documents.

1. Ignoring case, extract all unique words from the entire set of documents. CS 378 Introduction to Data Mining Spring 29 Lecture 2 Lecturer: Inderjit Dhillon Date: Jan. 27th, 29 Keywords: Vector space model, Latent Semantic Indexing(LSI), SVD 1 Vector Space Model The basic idea

More information

Image Registration Lecture 2: Vectors and Matrices

Image Registration Lecture 2: Vectors and Matrices Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this

More information

Using Matrix Decompositions in Formal Concept Analysis

Using Matrix Decompositions in Formal Concept Analysis Using Matrix Decompositions in Formal Concept Analysis Vaclav Snasel 1, Petr Gajdos 1, Hussam M. Dahwa Abdulla 1, Martin Polovincak 1 1 Dept. of Computer Science, Faculty of Electrical Engineering and

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

Large Scale Data Analysis Using Deep Learning

Large Scale Data Analysis Using Deep Learning Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu How to organize/navigate it? First try: Human curated Web directories Yahoo, DMOZ, LookSmart

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Talk 2: Graph Mining Tools - SVD, ranking, proximity. Christos Faloutsos CMU

Talk 2: Graph Mining Tools - SVD, ranking, proximity. Christos Faloutsos CMU Talk 2: Graph Mining Tools - SVD, ranking, proximity Christos Faloutsos CMU Outline Introduction Motivation Task 1: Node importance Task 2: Recommendations Task 3: Connection sub-graphs Conclusions Lipari

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed

More information

Wafer Pattern Recognition Using Tucker Decomposition

Wafer Pattern Recognition Using Tucker Decomposition Wafer Pattern Recognition Using Tucker Decomposition Ahmed Wahba, Li-C. Wang, Zheng Zhang UC Santa Barbara Nik Sumikawa NXP Semiconductors Abstract In production test data analytics, it is often that an

More information

Third-Order Tensor Decompositions and Their Application in Quantum Chemistry

Third-Order Tensor Decompositions and Their Application in Quantum Chemistry Third-Order Tensor Decompositions and Their Application in Quantum Chemistry Tyler Ueltschi University of Puget SoundTacoma, Washington, USA tueltschi@pugetsound.edu April 14, 2014 1 Introduction A tensor

More information

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:

More information

How does Google rank webpages?

How does Google rank webpages? Linear Algebra Spring 016 How does Google rank webpages? Dept. of Internet and Multimedia Eng. Konkuk University leehw@konkuk.ac.kr 1 Background on search engines Outline HITS algorithm (Jon Kleinberg)

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The

More information

Quick Introduction to Nonnegative Matrix Factorization

Quick Introduction to Nonnegative Matrix Factorization Quick Introduction to Nonnegative Matrix Factorization Norm Matloff University of California at Davis 1 The Goal Given an u v matrix A with nonnegative elements, we wish to find nonnegative, rank-k matrices

More information

Sparseness Constraints on Nonnegative Tensor Decomposition

Sparseness Constraints on Nonnegative Tensor Decomposition Sparseness Constraints on Nonnegative Tensor Decomposition Na Li nali@clarksonedu Carmeliza Navasca cnavasca@clarksonedu Department of Mathematics Clarkson University Potsdam, New York 3699, USA Department

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

Kronecker Product Approximation with Multiple Factor Matrices via the Tensor Product Algorithm

Kronecker Product Approximation with Multiple Factor Matrices via the Tensor Product Algorithm Kronecker Product Approximation with Multiple actor Matrices via the Tensor Product Algorithm King Keung Wu, Yeung Yam, Helen Meng and Mehran Mesbahi Department of Mechanical and Automation Engineering,

More information

CS 340 Lec. 6: Linear Dimensionality Reduction

CS 340 Lec. 6: Linear Dimensionality Reduction CS 340 Lec. 6: Linear Dimensionality Reduction AD January 2011 AD () January 2011 1 / 46 Linear Dimensionality Reduction Introduction & Motivation Brief Review of Linear Algebra Principal Component Analysis

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

Big Tensor Data Reduction

Big Tensor Data Reduction Big Tensor Data Reduction Nikos Sidiropoulos Dept. ECE University of Minnesota NSF/ECCS Big Data, 3/21/2013 Nikos Sidiropoulos Dept. ECE University of Minnesota () Big Tensor Data Reduction NSF/ECCS Big

More information

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology M. Soleymani Fall 2014 Most slides have been adapted from: Profs. Manning, Nayak & Raghavan (CS-276,

More information

Text Analytics (Text Mining)

Text Analytics (Text Mining) http://poloclub.gatech.edu/cse6242 CSE6242 / CX4242: Data & Visual Analytics Text Analytics (Text Mining) Concepts, Algorithms, LSI/SVD Duen Horng (Polo) Chau Assistant Professor Associate Director, MS

More information

Learning representations

Learning representations Learning representations Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 4/11/2016 General problem For a dataset of n signals X := [ x 1 x

More information

Postgraduate Course Signal Processing for Big Data (MSc)

Postgraduate Course Signal Processing for Big Data (MSc) Postgraduate Course Signal Processing for Big Data (MSc) Jesús Gustavo Cuevas del Río E-mail: gustavo.cuevas@upm.es Work Phone: +34 91 549 57 00 Ext: 4039 Course Description Instructor Information Course

More information

Matrix Decomposition in Privacy-Preserving Data Mining JUN ZHANG DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF KENTUCKY

Matrix Decomposition in Privacy-Preserving Data Mining JUN ZHANG DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF KENTUCKY Matrix Decomposition in Privacy-Preserving Data Mining JUN ZHANG DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF KENTUCKY OUTLINE Why We Need Matrix Decomposition SVD (Singular Value Decomposition) NMF (Nonnegative

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University.

CS246: Mining Massive Datasets Jure Leskovec, Stanford University. CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu What is the structure of the Web? How is it organized? 2/7/2011 Jure Leskovec, Stanford C246: Mining Massive

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu March 16, 2016 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision

More information

As it is not necessarily possible to satisfy this equation, we just ask for a solution to the more general equation

As it is not necessarily possible to satisfy this equation, we just ask for a solution to the more general equation Graphs and Networks Page 1 Lecture 2, Ranking 1 Tuesday, September 12, 2006 1:14 PM I. II. I. How search engines work: a. Crawl the web, creating a database b. Answer query somehow, e.g. grep. (ex. Funk

More information

Hyperlinked-Induced Topic Search (HITS) identifies. authorities as good content sources (~high indegree) HITS [Kleinberg 99] considers a web page

Hyperlinked-Induced Topic Search (HITS) identifies. authorities as good content sources (~high indegree) HITS [Kleinberg 99] considers a web page IV.3 HITS Hyperlinked-Induced Topic Search (HITS) identifies authorities as good content sources (~high indegree) hubs as good link sources (~high outdegree) HITS [Kleinberg 99] considers a web page a

More information

Slide source: Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University.

Slide source: Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University. Slide source: Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University http://www.mmds.org #1: C4.5 Decision Tree - Classification (61 votes) #2: K-Means - Clustering

More information