Extracting Optimal Performance from Dynamic Time Warping

Size: px
Start display at page:

Download "Extracting Optimal Performance from Dynamic Time Warping"

Transcription

1 Extracting Optimal Performance from Dynamic Time Warping Abdullah Mueen Eamonn Keogh While you are waiting: Please download the slides at: or../dtw1.pptx and../dtw2.pdf / DTW2.pptx

2 Structure There will be Q/A break before coffee break and another Q/A session at the end Interrupt only if you need clarification This segment has many high level concepts without minute details, follow embedded references for more information Items in many slides are color mapped, match colors for better understanding

3 The Second Act: How to do DTW fast We are motivated that DTW is GOOD by the first act The general conception: DTW is slow and we have a never-ending need for speed Better performance in knowledge extraction Better scalability to process BigData Better interactivity in human driven data analysis

4 What can be made fast? One-to-One comparison Exact Implementation and Constraints Efficient Approximation Exploiting Sparsity One-to-Many comparisons Nearest Neighbor Search In a database of independent time series In subsequences of a long time series Density Estimation In clustering Averaging Under Warping In classification Many-to-Many comparisons All-pair Distance Calculations

5 Speeding up DTW: one-to-one One-to-One comparison Exact Implementation Efficient Constraints Exploiting Hardware Efficient Approximation Exploiting Sparsity

6 Simplest Exact Implementation Input: x and y are time series of length n and m Output: DTW distance d between x and y O(n 2 ) time D(1:n+1,1:m+1) = inf; D(1,1) = 0; O(n 2 ) space for i = 2 : n+1 %for each row for j = 2 : m+1 %for each column cost = (x(i-1)-y(j-1))^2; D(i,j) = cost + min( [ D(i-1,j), D(i,j-1), D(i-1,j-1) ]); d = sqrt(d(n+1,m+1));

7 Simplest Implementation (Constrained) O(nw) time D(1:n+1,1:m+1) = inf; D(1,1) = 0; O(n 2 ) space w = max(w, abs(n-m)); for i = 2 : n+1 for j = max(2,i-w) : min(m+1,i+w) cost = (x(i-1)-y(j-1))^2; D(i,j) = cost + min( [ D(i-1,j), D(i,j-1), D(i-1,j-1) ]); d = sqrt(d(n+1,m+1)); w

8 Memoization O(n 2 ) time D(2,1:m+1) = inf; D(1,1) = 0; p = 1; c = 2; O(n) space for i = 2 : n+1 for j = 2 : m+1 cost = (x(i-1)-y(j-1))^2; D(c,j) = cost + min( [ D(p,j), D(c,j-1), D(p,j-1) ]); swap(c,p); d = sqrt(d(n+1,m+1)); Previous Row Current Row

9 Hardware Acceleration Single Instruction Multiple Data (SIMD) architecture Cells on a diagonal are computed in parallel Values of a diagonal depend on the previous two diagonals O(n) time O(n) space

10 PAA based approximation 2 time O n w 2 space O n w w Piecewise Aggregate Approximation Selina Chu, Eamonn J. Keogh, David M. Hart, Michael J. Pazzani: Iterative Deepening Dynamic Time Warping for Time Series. SDM 2002:

11 Approximation by Lengthencoding To exploit sparsity, encode lengths of the runs of zeros 1 (4) 1 (4) 1 1 (3) 1 (3) 1 1 (3) 1 A Mueen, N Chavoshi, N Abu-El-Rub, H Hamooni, A Minnich, Fast Warping Distance for Sparse Time Series, Technical Report at UNM:

12 Exploiting Sparsity (1) y x

13 Exploiting Sparsity (2) y x O(nm) time and space X Y 1 (3) 1 (3) 1 1 (3) 1 1 (4) 1 (4) O(NM) time and space

14 Exploiting Sparsity (2) Correct Alignment Missing Alignment Extra Alignment Non-Linear change Exact Distance y x y No change Lower bound (1) 1 x 1 (2) y x Linear change Upper bound (1) 1 1 (2)

15 Percentage Percentage Percentage Percentage Exploiting Sparsity (3) Awarp_UB 1.05 * DTW AWarp_UB = DTW AWarp_LB = DTW Normal Distribution Sparsity Factor Exponential Distribution Awarp_UB 1.05 * DTW AWarp_UB = DTW AWarp_LB = DTW Sparsity Factor Awarp_UB 1.05 * DTW AWarp_UB = DTW AWarp_LB = DTW Uniform Distribution Sparsity Factor Awarp_UB 1.05 * DTW AWarp_UB = DTW AWarp_LB = DTW Binomial Distribution Sparsity Factor Sparsity Factor of s means 1 % of the time s series is filled with nonzeros.

16 What can be made fast? One-to-One comparison Exact Implementation and Constraints Efficient Approximation Exploiting Sparsity One-to-Many comparisons Nearest Neighbor Search In a database of independent time series In subsequences of a long time series Density Estimation In clustering Averaging Under Warping In classification Many-to-Many comparisons All-pair Distance Calculations

17 Nearest Neighbor Search A query Q is given n independent candidate time series C 1, C 2,, C n O(n) distance calculations are performed to Find THE nearest neighbor of the given query under DTW.

18 Brute Force Nearest Neighbor Search Computational cost: O(nm 2 ) Algorithm Sequential_Scan(Q) 1. best_so_far = infinity; 2. for all sequences in database true_dist = DTW(C i,q); 6. if true_dist < best_so_far 7. best_so_far = true_dist; 8. index_of_best_match = i; 9. endif endfor

19 Lower Bounding Nearest Neighbor Search We can speed up similarity search under DTW by using a lower bounding function Algorithm Lower_Bounding_Sequential_Scan(Q) 1. best_so_far = infinity; 2. for all sequences in database LB_dist = lower_bound_distance(c i,q); if LB_dist < best_so_far true_dist = DTW(C i,q); if true_dist < best_so_far 7. best_so_far = true_dist; 8. index_of_best_match = i; 9. endif 10. endif 11. endfor Try to use a cheap lower bounding calculation as often as possible. Only do the expensive, full calculations when it is absolutely necessary

20 Lower Bound of Kim C A D LB_Kim B O(1) time if considered only first and last points O(n) time for all four distances Kim, S, Park, S, & Chu, W. An index-based approach for similarity search supporting time warping in large sequence databases. ICDE 01, pp The squared difference between the two sequence s first (A), last (D), minimum (B) and maximum points (C) is returned as the lower bound

21 Lower Bound of Yi max(q) LB_Yi O(n) time Yi, B, Jagadish, H & Faloutsos, C. Efficient retrieval of similar time sequences under time warping. ICDE 98, pp min(q) The sum of the squared length of white lines represent the minimum contribution of the observations above and below the yellow lines.

22 Q Lower Bound of Keogh C U i = max(q i-w : q i+w ) L i = min(q i-w : q i+w ) U Sakoe-Chiba Band O(n) time Q L Envelope-Based Lower Bound U C LB_ Keogh( Q, C) n i 1 ( q i U ) i 2 if q U 2 ( qi Li ) if qi Li 0 otherwise i i Q L

23 Reversing the Query/Data Role in LB_Keogh Make LB_Keogh tighter Much cheaper than DTW U/L envelops on the candidates can be calculated online or pre-calculated max(lb _Keogh EQ, LB _Keogh EC) Envelop on Q Envelop on C U C U C Q L Q L 24

24 The tightness of the lower bound for each technique is proportional to the length of lines used in the illustrations LB_Kim LB_Yi LB_Keogh Sakoe-Chiba LB_Keogh Itakura

25 Tightness of lower bound Cascading Lower Bounds At least 18 lower bounds of DTW was proposed. Use lower bounds only on the Skyline. Use the bounds on the skyline in cascade from least expensive to most expensive When unable to prune, use early abandoning techniques 1 0 LB_KimFL Early_abandoning_DTW max(lb_ Keogh EQ, LB_ Keogh EC) LB_ Keogh EQ LB_Yi LB_Kim LB_FTW LB_Ecorner LB_PAA O(1) O(n) O(nR) DTW 99.9% of the time DTW is not calculated 26

26 Early Abandoning Techniques Abandon accumulating errors as soon as the current total is larger than the best_so_far Four techniques to abandon early 1. Early Abandoning of LB_Keogh 2. Early Abandoning of DTW 3. Earlier Early Abandoning of DTW using LB_Keogh 4. Reordering Early Abandoning

27 Early Abandoning of LB_Keogh Abandon the computation, when the accumulated error is larger than best_so_far U C Q L U, L are upper and lower envelopes of Q 28

28 Early Abandoning of DTW Q C Abandon the computation, when the dtw_dist is larger than best_so_far C dtw_dist R (Warping Windows) Q 29

29 Earlier Early Abandoning of DTW using LB_Keogh Abandon the computation, when the dtw_dist + lb_keogh is larger than best_so_far C Q (partial) lb_keogh U C R (Warping Windows) (partial) dtw_dist Q L U, L are upper and lower envelopes of Q 30

30 Reordering Early Abandoning We don t have to compute LB from left to right. Order points by expected contribution U C Idea Q L - Order by the absolute height of the query point. 31

31 Summary of the techniques Group-1 Techniques Early Abandoning of LB_Keogh Early Abandoning of DTW Earlier Early Abandoning of DTW using LB_Keogh Group-2 Techniques Just-in-time Z-normalizations Reordering Early Abandoning Reversing LB_Koegh Cascading Lower Bounds UCR Suite Code and data is available at: Thanawin Rakthanmanon, Bilson J. L. Campana, Abdullah Mueen, Gustavo E. A. P. A. Batista, M. Brandon Westover, Qiang Zhu, Jesin Zakaria, Eamonn J. Keogh: Searching and mining trillions of time series subsequences under dynamic time warping. KDD 2012:

32 Experimental Result: Random Walk Random Walk: Varying size of the data Million (Seconds) Billion (Minutes) Trillion (Hours) DTW-Naive , ,869 Group Group-1 and

33 Experimental Result: Random Walk Random Walk: Varying size of the query seconds For query lengths of 4,096 (rightmost part of this graph) The times are: Naïve DTW 24,286 Group-1 5,078 Group-1 and Naïve DTW Group-1 Techniques Group 1 and Query Length Power of two 34

34 Seconds Experimental Result: Random Walk Random Walk: Varying size of the band 600 DTW-Naïve Group-1 Group-1 and Sakoe-Chiba Band Width (percentage of query length)

35 Nearest Subsequence Search A Q query is given A long time series of length n O(n) distance calculations are performed to Find THE nearest subsequence of the given query under DTW.

36 Time Warping Subsequence Search S Reuses computation for subsequence matching Q Match 1 Match 2 Match 3 Match 4 For every new observation only one column is added on the right No need for any of the techniques 37 Yasushi Sakurai, Christos Faloutsos, Masashi Yamamuro: Stream Monitoring under the Time Warping Distance. ICDE 2007:

37 Normalization is required If each window is normalized separately, reuse of computation is no longer possible To take advantage of the bounding and abandoning techniques, we need just-in-time normalization with constant overhead per comparison

38 Just-in-time Normalization In one pass, calculate cumulative sums of over x and x 2 and store C = x C 2 = x 2 Subtract two cumulative sums to obtain the sum over a window S S i = C i+w C i i = C i+w C i Use the sums to calculate the means and standard deviations of all windows in linear time μ i = S i w Dynamically normalize observations when calculating distance and possibly abandon early cost = σ i = S i 2 x ij μ xi σ xi w S i w q i μ qi σ qi

39 Experimental Result: ECG Data: One year of Electrocardiograms 8.5 billion data points. Query: Idealized Premature Ventricular Contraction (PVC) of length 421 (R=21=5%). PVC (aka. skipped beat) Group-1 Group 1 & 2 ECG 49.2 hours 18.0 minutes ~30,000X faster than real time! 40

40 Experimental Result: DNA Query: Human Chromosome 2 of length 72,500 bps Data: Chimp Genome 2.9 billion bps Time: UCR Suite 14.6 hours, SOTA 34.6 days (830 hours) Rhesus macaque Gibbon Orangutan Gorilla Chimp Human Catarrhines Hominoidea Hominidae Homininae Hominini Chromosome 2: BP :

41 What can be made fast? One-to-One comparison Exact Implementation and Constraints Efficient Approximation Exploiting Sparsity One-to-Many comparisons Nearest Neighbor Search In a database of independent time series In subsequences of a long time series Density Estimation In clustering Averaging Under Warping In classification Many-to-Many comparisons All-pair Distance Calculations

42 Density based clustering Density Peaks (DP)* Algorithm Find the densities of every point to pick cluster centers Connect every point to the nearest higher density point *Rodriguez, A., & Laio, A. (2014). Clustering by fast search and find of density peaks. Science, 344(6191),

43 Range Search/Density Estimation Density is estimated by the number of points within a radius/threshold t Algorithm Bounding_Range_Search(Q,t) 1. for all sequences C i in database 2. LB_dist = lower_bound_distance (C i, Q) 3. if LB_dist < t 4. UB_dist = upper_bound_distance (C i, Q) if UB_dist < t then output C i else true_dist = DTW(C i, Q) if true_dist < t output C i 10. endif 11. endif 12. endif 13. endfor Try to use an upper bound to identify a point within the range Nurjahan Begum, Liudmila Ulanova, Jun Wang, Eamonn J. Keogh: Accelerating Dynamic Time Warping Clustering with a Novel Admissible Pruning Strategy. KDD 2015: 49-58

44 Density Connectedness Distance between a pair of points is an upper bound of the NN distance from both of the points Algorithm Bounding_Scan(D,Q) 1. best_so_far =min(upper_bound_nn_distance(d,q) 2. for all sequences in D 3. LB_dist = lower_bound_distance( C i, Q); 4. if LB_dist < best_so_far 5. true_dist = DTW( C i, Q); 6. if true_dist < best_so_far 7. best_so_far = true_dist; 8. index_of_best_match = i; Try to use an 9. endif upper bound to the 10. endif NN distance as 11. endfor the best_so_far Q D Nurjahan Begum, Liudmila Ulanova, Jun Wang, Eamonn J. Keogh: Accelerating Dynamic Time Warping Clustering with a Novel Admissible Pruning Strategy. KDD 2015: 49-58

45 DTW Distance Upper bounding Euclidean distance is a trivial upper bound DTW distance in a band w is an upper bound for DTW distance in band w Zoom-In Euclidean Distance

46 Distance Calculations Distance Calculations Speedup by upper bounds Density Peak: 9 Hours TADPole: 9 minutes 7 5 x 10 6 Absolute Number 100 Brute force Percentages 3 1 TADPole 0 Number of objects 3500 TADPole 0 0 Number of objects 3500 StarLightCurves dataset 47

47 What can be made fast? One-to-One comparison Exact Implementation and Constraints Efficient Approximation Exploiting Sparsity One-to-Many comparisons Nearest Neighbor Search In a database of independent time series In subsequences of a long time series Density Estimation In clustering Averaging Under Warping In classification Many-to-Many comparisons All-pair Distance Calculations

48 Data Reduction for 1NN Classification The training set is reduced to a smaller set keeping a representative set of labeled instances Smaller training set entails performance gain Smaller training set may gain accuracy if noisy instances are filtered effectively Reduction methods Random Selection Rank the instances and take top-k Cluster instances based on proximity and take representative from each cluster

49 Many clustering algorithms require finding a centroid of two or more instances The issue is then: How to average time series consistently with DTW? Compute average Trace dataset François Petitjean, Germain Forestier, Geoffrey I. Webb, Ann E. Nicholson, Yanping Chen, Eamonn J. Keogh: Dynamic Time Warping Averaging of Time Series Allows Faster and More Accurate Classification. ICDM 2014:

50 Mathematically, the mean oҧ of a set of objects O embedded in a space induced by a distance d is: arg min തo d 2 o, ҧ o o O The mean of a set minimizes the sum of the squared distances If d is the Euclidean distance The arithmetic mean solves the problem exactly o ҧ = 1 N o o O Arithmetic mean 51

51 To solve the optimization problem for DTW distance, we need to perform simultaneous alignment of many time series. But, finding the optimal multiple alignment: 1. Is NP-complete [a] 2. Requires O L N operations L is the length of the sequences ( 100) N is the number of sequences ( 1,000) Efficient solutions will be heuristic Pairwise Averaging DTW Barycenter Averaging (DBA) [a] F. Petitjean, A. Ketterlin and P. Gançarski, A global averaging method for dynamic time warping, with applications to clustering, Pattern Recognition, vol. 44, no. 3, pp , 2011.

52 Pairwise averaging for DTW Average each alignment between the two time series Commonly increases the length Chaining can produce average over a set The operation is not associative, the average produced depends on the order Average X 1 X 2 Average X 1,2 Average X 1-4 X 3 Average X 3,4 X 4 V. Niennattrakul and C. A. Ratanamahatana, On Clustering Multimedia Time Series Data Using K-Means and Dynamic Time Warping, IEEE International Conference on Multimedia and Ubiquitous Engineering, pp , 2007.

53 DTW Barycenter Averaging (DBA) Algorithm DBA(D,av) 1 Iterate until convergence 2 for each series s i in D 3 A i = GetAlignment(DTW(s i, av)) 4 for each sample j in av 5 av[j] = mean([a 1 [j] A 2 [j] A 3 [j]. A n [j]) av s 1 s 1 s 2 s 2

54 Error-Rate Experimental Evaluation on Insect Data 0.3 SR Drop1 Drop3 random Drop2 4 rank-based competitors 1. Drop 1 2. Drop 2 3. Drop 3 4. Simple Rank 2 average-based techniques 1. K-means 2. AHC both using DBA 0.2 AHC Kmeans 0.1 The minimum error-rate is 0.092, with 19 pairs of objects The full dataset error-rate is 0.14, with 100 pairs of objects Items per class in reduced training set Code Available : 55

55 What can be made fast? One-to-One comparison Exact Implementation and Constraints Efficient Approximation Exploiting Sparsity One-to-Many comparisons Nearest Neighbor Search In a database of independent time series In subsequences of a long time series Density Estimation In clustering Averaging Under Warping In classification Many-to-Many comparisons All-pair Distance Calculations

56 [a] N Chavoshi, H Hamooni, A Mueen, DeBot: Real-Time Bot Detection via Activity Correlation UNM Technical Report [b] Y. Chen, G. Chen, K. Chen and B. C. Ooi, "Efficient Processing of Warping Time Series Join of Motion Capture Data," ICDE, 2009, pp Speeding up DTW: many-to-many Several variants Self-join within a threshold - top-k Self-join Use similarity search techniques as subroutine Application: Motif discovery [a], Discord discovery A/B Join within a threshold - top-k A/B Join Use similarity search techniques as subroutine Application: Motion Stitching [b] All-pair distance matrix Use techniques to speedup one-to-one comparisons Application: Hierarchical Clustering

57 PrunedDTW: speeding up all-pair distance matrix calculation Two types of pruning when calculating DTW matrix UB >UB >UB UB UB UB Exact method 6 UB Lower triangle pruning UB UB UB UB UB UB >UB >UB >UB >UB UB = Euclidean distance UB = DTW distance Upper triangle pruning Diego F. Silva, Gustavo E. A. P. A. Batista, Speeding Up All-Pairwise Dynamic Time Warping Matrix Calculation, SDM 2016

58 Time (s) Experiments x Car CinC ECG torso Haptics InlineSkate Lighting-2 x x x x MALLAT Non-Invasive Fetal ECG Non-Invasive Fetal ECG Olive Oil Starlight Curves DTW Pruned DTW Oracle DTW Warping Window Length

59 Conclusion Nearest neighbor search under warping is fast enough for most practical purposes New invariances (e.g. normalization) lead to challenging problems Data reduction improves 1NN classification both in speed and accuracy DTW is an extraordinarily powerful and useful tool. Its uses are limited only by our imagination.

60

Efficient search of the best warping window for Dynamic Time Warping

Efficient search of the best warping window for Dynamic Time Warping Efficient search of the best warping window for Dynamic Time Warping Chang Wei Tan 1 Matthieu Herrmann 1 Germain Geoffrey I. Forestier 2,1 Webb 1 François Petitjean 1 1 Faculty of IT, Monash University,

More information

Dynamic Time Warping

Dynamic Time Warping Dynamic Time Warping Steven Elsworth School of Mathematics, University of Manchester Friday 29 th September, 2017 S. Elsworth Dynamic Time Warping 1 / 19 What is a time series? A time series is a sequence

More information

Time-Series Analysis Prediction Similarity between Time-series Symbolic Approximation SAX References. Time-Series Streams

Time-Series Analysis Prediction Similarity between Time-series Symbolic Approximation SAX References. Time-Series Streams Time-Series Streams João Gama LIAAD-INESC Porto, University of Porto, Portugal jgama@fep.up.pt 1 Time-Series Analysis 2 Prediction Filters Neural Nets 3 Similarity between Time-series Euclidean Distance

More information

Outline. Background on time series mining Similarity Measures Normalization

Outline. Background on time series mining Similarity Measures Normalization Reconvene In the first half, we have seen many wonderful things you can do with the Matrix Profile, without explaining how to compute it! In this half, we will describe algorithms to compute matrix profile,

More information

Anticipatory DTW for Efficient Similarity Search in Time Series Databases

Anticipatory DTW for Efficient Similarity Search in Time Series Databases Anticipatory DTW for Efficient Similarity Search in Time Series Databases Ira Assent Marc Wichterich Ralph Krieger Hardy Kremer Thomas Seidl RWTH Aachen University, Germany Aalborg University, Denmark

More information

HOT SAX: Efficiently Finding the Most Unusual Time Series Subsequence

HOT SAX: Efficiently Finding the Most Unusual Time Series Subsequence HOT SAX: Efficiently Finding the Most Unusual Time Series Subsequence Eamonn Keogh *Jessica Lin Ada Fu University of California, Riverside George Mason Univ Chinese Univ of Hong Kong The 5 th IEEE International

More information

Composite Quantization for Approximate Nearest Neighbor Search

Composite Quantization for Approximate Nearest Neighbor Search Composite Quantization for Approximate Nearest Neighbor Search Jingdong Wang Lead Researcher Microsoft Research http://research.microsoft.com/~jingdw ICML 104, joint work with my interns Ting Zhang from

More information

Time Series Classification

Time Series Classification Distance Measures Classifiers DTW vs. ED Further Work Questions August 31, 2017 Distance Measures Classifiers DTW vs. ED Further Work Questions Outline 1 2 Distance Measures 3 Classifiers 4 DTW vs. ED

More information

Learning Time-Series Shapelets

Learning Time-Series Shapelets Learning Time-Series Shapelets Josif Grabocka, Nicolas Schilling, Martin Wistuba and Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) University of Hildesheim, Germany SIGKDD 14,

More information

Low redundancy estimation of correlation matrices for time series using triangular bounds

Low redundancy estimation of correlation matrices for time series using triangular bounds Low redundancy estimation of correlation matrices for time series using triangular bounds Erik Scharwächter 1,2, Fabian Geier 2, Lukas Faber 2, and Emmanuel Müller 1,2 1 GFZ German Research Centre for

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu March 16, 2016 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu November 16, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision

More information

Improving Performance of Similarity Measures for Uncertain Time Series using Preprocessing Techniques

Improving Performance of Similarity Measures for Uncertain Time Series using Preprocessing Techniques Improving Performance of Similarity Measures for Uncertain Time Series using Preprocessing Techniques Mahsa Orang Nematollaah Shiri 27th International Conference on Scientific and Statistical Database

More information

Improving Performance of Similarity Measures for Uncertain Time Series using Preprocessing Techniques

Improving Performance of Similarity Measures for Uncertain Time Series using Preprocessing Techniques Improving Performance of Similarity Measures for Uncertain Time Series using Preprocessing Techniques Mahsa Orang Nematollaah Shiri 27th International Conference on Scientific and Statistical Database

More information

Time Series Shapelets

Time Series Shapelets Time Series Shapelets Lexiang Ye and Eamonn Keogh This file contains augmented versions of the figures in our paper, plus additional experiments and details that were omitted due to space limitations Please

More information

On-line Dynamic Time Warping for Streaming Time Series

On-line Dynamic Time Warping for Streaming Time Series On-line Dynamic Time Warping for Streaming Time Series Izaskun Oregi 1, Aritz Pérez 2, Javier Del Ser 1,2,3, and José A. Lozano 2,4 1 TECNALIA, 48160 Derio, Spain {izaskun.oregui,javier.delser}@tecnalia.com

More information

Fusion of Similarity Measures for Time Series Classification

Fusion of Similarity Measures for Time Series Classification Fusion of Similarity Measures for Time Series Classification Krisztian Buza, Alexandros Nanopoulos, and Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) University of Hildesheim,

More information

Identifying Patterns and Anomalies in Delayed Neutron Monitor Data of Nuclear Power Plant

Identifying Patterns and Anomalies in Delayed Neutron Monitor Data of Nuclear Power Plant Identifying Patterns and Anomalies in Delayed Neutron Monitor Data of Nuclear Power Plant Durga Toshniwal, Aditya Gupta Department of Computer Science & Engineering Indian Institute of Technology Roorkee

More information

Coarse-DTW for Sparse Time Series Alignment

Coarse-DTW for Sparse Time Series Alignment Coarse-DTW for Sparse Time Series Alignment Marc Dupont 1,2 and Pierre-François Marteau 1 1 IRISA, Université de Bretagne Sud, Campus de Tohannic, Vannes, France 2 Thales Optronique, 2 Avenue Gay Lussac,

More information

Machine Learning on temporal data

Machine Learning on temporal data Machine Learning on temporal data Learning Dissimilarities on Time Series Ahlame Douzal (Ahlame.Douzal@imag.fr) AMA, LIG, Université Joseph Fourier Master 2R - MOSIG (2011) Plan Time series structure and

More information

Università di Pisa A.A Data Mining II June 13th, < {A} {B,F} {E} {A,B} {A,C,D} {F} {B,E} {C,D} > t=0 t=1 t=2 t=3 t=4 t=5 t=6 t=7

Università di Pisa A.A Data Mining II June 13th, < {A} {B,F} {E} {A,B} {A,C,D} {F} {B,E} {C,D} > t=0 t=1 t=2 t=3 t=4 t=5 t=6 t=7 Università di Pisa A.A. 2016-2017 Data Mining II June 13th, 2017 Exercise 1 - Sequential patterns (6 points) a) (3 points) Given the following input sequence < {A} {B,F} {E} {A,B} {A,C,D} {F} {B,E} {C,D}

More information

Generalized Gradient Learning on Time Series under Elastic Transformations

Generalized Gradient Learning on Time Series under Elastic Transformations Generalized Gradient Learning on Time Series under Elastic Transformations Brijnesh J. Jain Technische Universität Berlin, Germany e-mail: brijnesh.jain@gmail.com arxiv:1502.04843v2 [cs.lg] 9 Jun 2015

More information

Estimating the Selectivity of tf-idf based Cosine Similarity Predicates

Estimating the Selectivity of tf-idf based Cosine Similarity Predicates Estimating the Selectivity of tf-idf based Cosine Similarity Predicates Sandeep Tata Jignesh M. Patel Department of Electrical Engineering and Computer Science University of Michigan 22 Hayward Street,

More information

Temporal and Frequential Metric Learning for Time Series knn Classication

Temporal and Frequential Metric Learning for Time Series knn Classication Proceedings 1st International Workshop on Advanced Analytics and Learning on Temporal Data AALTD 2015 Temporal and Frequential Metric Learning for Time Series knn Classication Cao-Tri Do 123, Ahlame Douzal-Chouakria

More information

arxiv: v2 [cs.lg] 13 Mar 2018

arxiv: v2 [cs.lg] 13 Mar 2018 arxiv:1802.03628v2 [cs.lg] 13 Mar 2018 ABSTRACT Han Qiu Massachusetts Institute of Technology Cambridge, MA, USA hanqiu@mit.edu Francesco Fusco IBM Research Dublin, Ireland francfus@ie.ibm.com We propose

More information

Generalized gradient learning on time series

Generalized gradient learning on time series Mach Learn (2015) 100:587 608 DOI 10.1007/s10994-015-5513-0 Generalized gradient learning on time series Brijnesh J. Jain 1 Received: 18 January 2015 / Accepted: 3 June 2015 / Published online: 16 July

More information

2.6 Complexity Theory for Map-Reduce. Star Joins 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51

2.6 Complexity Theory for Map-Reduce. Star Joins 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51 Star Joins A common structure for data mining of commercial data is the star join. For example, a chain store like Walmart keeps a fact table whose tuples each

More information

ECS289: Scalable Machine Learning

ECS289: Scalable Machine Learning ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Oct 18, 2016 Outline One versus all/one versus one Ranking loss for multiclass/multilabel classification Scaling to millions of labels Multiclass

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu Find most influential set S of size k: largest expected cascade size f(s) if set S is activated

More information

ECE 661: Homework 10 Fall 2014

ECE 661: Homework 10 Fall 2014 ECE 661: Homework 10 Fall 2014 This homework consists of the following two parts: (1) Face recognition with PCA and LDA for dimensionality reduction and the nearest-neighborhood rule for classification;

More information

Slides credits: Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University

Slides credits: Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University Note to other teachers and users of these slides: We would be delighted if you found this our material useful in giving your own lectures. Feel free to use these slides verbatim, or to modify them to fit

More information

Dimension reduction methods: Algorithms and Applications Yousef Saad Department of Computer Science and Engineering University of Minnesota

Dimension reduction methods: Algorithms and Applications Yousef Saad Department of Computer Science and Engineering University of Minnesota Dimension reduction methods: Algorithms and Applications Yousef Saad Department of Computer Science and Engineering University of Minnesota Université du Littoral- Calais July 11, 16 First..... to the

More information

Rare Time Series Motif Discovery from Unbounded Streams

Rare Time Series Motif Discovery from Unbounded Streams Rare Time Series Motif Discovery from Unbounded Streams Nurjahan Begum UC Riverside nbegu@cs.ucr.edu Eamonn Keogh UC Riverside eamonn@cs.ucr.edu ABSTRACT The detection of time series motifs, which are

More information

Greedy Maximization Framework for Graph-based Influence Functions

Greedy Maximization Framework for Graph-based Influence Functions Greedy Maximization Framework for Graph-based Influence Functions Edith Cohen Google Research Tel Aviv University HotWeb '16 1 Large Graphs Model relations/interactions (edges) between entities (nodes)

More information

CS249: ADVANCED DATA MINING

CS249: ADVANCED DATA MINING CS249: ADVANCED DATA MINING Graph and Network Instructor: Yizhou Sun yzsun@cs.ucla.edu May 31, 2017 Methods Learnt Classification Clustering Vector Data Text Data Recommender System Decision Tree; Naïve

More information

Machine Learning. Nonparametric Methods. Space of ML Problems. Todo. Histograms. Instance-Based Learning (aka non-parametric methods)

Machine Learning. Nonparametric Methods. Space of ML Problems. Todo. Histograms. Instance-Based Learning (aka non-parametric methods) Machine Learning InstanceBased Learning (aka nonparametric methods) Supervised Learning Unsupervised Learning Reinforcement Learning Parametric Non parametric CSE 446 Machine Learning Daniel Weld March

More information

Dynamic Time Warping and FFT: A Data Preprocessing Method for Electrical Load Forecasting

Dynamic Time Warping and FFT: A Data Preprocessing Method for Electrical Load Forecasting Vol. 9, No., 18 Dynamic Time Warping and FFT: A Data Preprocessing Method for Electrical Load Forecasting Juan Huo School of Electrical and Automation Engineering Zhengzhou University, Henan Province,

More information

Recent advances in Time Series Classification

Recent advances in Time Series Classification Distance Shapelet BoW Kernels CCL Recent advances in Time Series Classification Simon Malinowski, LinkMedia Research Team Classification day #3 S. Malinowski Time Series Classification 21/06/17 1 / 55

More information

SBASS: Segment based approach for subsequence searches in sequence databases

SBASS: Segment based approach for subsequence searches in sequence databases Comput Syst Sci & Eng (2007) 1-2: 37-46 2007 CRL Publishing Ltd International Journal of Computer Systems Science & Engineering SBASS: Segment based approach for subsequence searches in sequence databases

More information

Scalable Algorithms for Distribution Search

Scalable Algorithms for Distribution Search Scalable Algorithms for Distribution Search Yasuko Matsubara (Kyoto University) Yasushi Sakurai (NTT Communication Science Labs) Masatoshi Yoshikawa (Kyoto University) 1 Introduction Main intuition and

More information

IBM Research Report. A Lower Bound on the Euclidean Distance for Fast Nearest Neighbor Retrieval in High-dimensional Spaces

IBM Research Report. A Lower Bound on the Euclidean Distance for Fast Nearest Neighbor Retrieval in High-dimensional Spaces RC24859 (W0909-03) September 0, 2009 Computer Science IBM Research Report A Lower Bound on the Euclidean Distance for Fast Nearest Neighbor Retrieval in High-dimensional Spaces George Saon, Peder Olsen

More information

Iterative Laplacian Score for Feature Selection

Iterative Laplacian Score for Feature Selection Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,

More information

Online Discovery of Group Level Events in Time Series

Online Discovery of Group Level Events in Time Series Online Discovery of Group Level Events in Time Series Abdullah Mueen chen@cs.umn.edu mueen@cs.unm.edu Xi C. Chen Vijay K Narayanan Nikos Karampatziakis vkn@microsoft.com nikosk@microsoft.com Gagan Bansal

More information

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication.

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication. CME342 Parallel Methods in Numerical Analysis Matrix Computation: Iterative Methods II Outline: CG & its parallelization. Sparse Matrix-vector Multiplication. 1 Basic iterative methods: Ax = b r = b Ax

More information

Time Series (1) Information Systems M

Time Series (1) Information Systems M Time Series () Information Systems M Prof. Paolo Ciaccia http://www-db.deis.unibo.it/courses/si-m/ Time series are everywhere Time series, that is, sequences of observations (samples) made through time,

More information

Nearest Neighbor Search for Relevance Feedback

Nearest Neighbor Search for Relevance Feedback Nearest Neighbor earch for Relevance Feedbac Jelena Tešić and B.. Manjunath Electrical and Computer Engineering Department University of California, anta Barbara, CA 93-9 {jelena, manj}@ece.ucsb.edu Abstract

More information

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan Clustering CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Supervised vs Unsupervised Learning Supervised learning Given x ", y " "%& ', learn a function f: X Y Categorical output classification

More information

CS60021: Scalable Data Mining. Large Scale Machine Learning

CS60021: Scalable Data Mining. Large Scale Machine Learning J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org 1 CS60021: Scalable Data Mining Large Scale Machine Learning Sourangshu Bhattacharya Example: Spam filtering Instance

More information

CSE446: non-parametric methods Spring 2017

CSE446: non-parametric methods Spring 2017 CSE446: non-parametric methods Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin and Luke Zettlemoyer Linear Regression: What can go wrong? What do we do if the bias is too strong? Might want

More information

A Bayesian Method for Guessing the Extreme Values in a Data Set

A Bayesian Method for Guessing the Extreme Values in a Data Set A Bayesian Method for Guessing the Extreme Values in a Data Set Mingxi Wu University of Florida May, 2008 Mingxi Wu (University of Florida) May, 2008 1 / 74 Outline Problem Definition Example Applications

More information

Notion of Distance. Metric Distance Binary Vector Distances Tangent Distance

Notion of Distance. Metric Distance Binary Vector Distances Tangent Distance Notion of Distance Metric Distance Binary Vector Distances Tangent Distance Distance Measures Many pattern recognition/data mining techniques are based on similarity measures between objects e.g., nearest-neighbor

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

A Representation of Time Series for Temporal Web Mining

A Representation of Time Series for Temporal Web Mining A Representation of Time Series for Temporal Web Mining Mireille Samia Institute of Computer Science Databases and Information Systems Heinrich-Heine-University Düsseldorf D-40225 Düsseldorf, Germany samia@cs.uni-duesseldorf.de

More information

Predictive analysis on Multivariate, Time Series datasets using Shapelets

Predictive analysis on Multivariate, Time Series datasets using Shapelets 1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,

More information

Tailored Bregman Ball Trees for Effective Nearest Neighbors

Tailored Bregman Ball Trees for Effective Nearest Neighbors Tailored Bregman Ball Trees for Effective Nearest Neighbors Frank Nielsen 1 Paolo Piro 2 Michel Barlaud 2 1 Ecole Polytechnique, LIX, Palaiseau, France 2 CNRS / University of Nice-Sophia Antipolis, Sophia

More information

Outline for today. Information Retrieval. Cosine similarity between query and document. tf-idf weighting

Outline for today. Information Retrieval. Cosine similarity between query and document. tf-idf weighting Outline for today Information Retrieval Efficient Scoring and Ranking Recap on ranked retrieval Jörg Tiedemann jorg.tiedemann@lingfil.uu.se Department of Linguistics and Philology Uppsala University Efficient

More information

EVOLUTIONARY DISTANCES

EVOLUTIONARY DISTANCES EVOLUTIONARY DISTANCES FROM STRINGS TO TREES Luca Bortolussi 1 1 Dipartimento di Matematica ed Informatica Università degli studi di Trieste luca@dmi.units.it Trieste, 14 th November 2007 OUTLINE 1 STRINGS:

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

Proximity-Based Anomaly Detection using Sparse Structure Learning

Proximity-Based Anomaly Detection using Sparse Structure Learning Proximity-Based Anomaly Detection using Sparse Structure Learning Tsuyoshi Idé (IBM Tokyo Research Lab) Aurelie C. Lozano, Naoki Abe, and Yan Liu (IBM T. J. Watson Research Center) 2009/04/ SDM 2009 /

More information

On the Projection Matrices Influence in the Classification of Compressed Sensed ECG Signals

On the Projection Matrices Influence in the Classification of Compressed Sensed ECG Signals On the Projection Matrices Influence in the Classification of Compressed Sensed ECG Signals Monica Fira, Liviu Goras Institute of Computer Science Romanian Academy Iasi, Romania Liviu Goras, Nicolae Cleju,

More information

CS 273 Prof. Serafim Batzoglou Prof. Jean-Claude Latombe Spring Lecture 12 : Energy maintenance (1) Lecturer: Prof. J.C.

CS 273 Prof. Serafim Batzoglou Prof. Jean-Claude Latombe Spring Lecture 12 : Energy maintenance (1) Lecturer: Prof. J.C. CS 273 Prof. Serafim Batzoglou Prof. Jean-Claude Latombe Spring 2006 Lecture 12 : Energy maintenance (1) Lecturer: Prof. J.C. Latombe Scribe: Neda Nategh How do you update the energy function during the

More information

Algorithms for Calculating Statistical Properties on Moving Points

Algorithms for Calculating Statistical Properties on Moving Points Algorithms for Calculating Statistical Properties on Moving Points Dissertation Proposal Sorelle Friedler Committee: David Mount (Chair), William Gasarch Samir Khuller, Amitabh Varshney January 14, 2009

More information

CS425: Algorithms for Web Scale Data

CS425: Algorithms for Web Scale Data CS: Algorithms for Web Scale Data Most of the slides are from the Mining of Massive Datasets book. These slides have been modified for CS. The original slides can be accessed at: www.mmds.org Customer

More information

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS Learning Neural Networks Classifier Short Presentation INPUT: classification data, i.e. it contains an classification (class) attribute.

More information

Metric-based classifiers. Nuno Vasconcelos UCSD

Metric-based classifiers. Nuno Vasconcelos UCSD Metric-based classifiers Nuno Vasconcelos UCSD Statistical learning goal: given a function f. y f and a collection of eample data-points, learn what the function f. is. this is called training. two major

More information

MIRA, SVM, k-nn. Lirong Xia

MIRA, SVM, k-nn. Lirong Xia MIRA, SVM, k-nn Lirong Xia Linear Classifiers (perceptrons) Inputs are feature values Each feature has a weight Sum is the activation activation w If the activation is: Positive: output +1 Negative, output

More information

Introduction to Time Series Mining. Slides edited from Keogh Eamonn s tutorial:

Introduction to Time Series Mining. Slides edited from Keogh Eamonn s tutorial: Introduction to Time Series Mining Slides edited from Keogh Eamonn s tutorial: 25.1750 25.2250 25.2500 25.2500 25.2750 25.3250 25.3500 25.3500 25.4000 25.4000 25.3250 25.2250 25.2000 25.1750.... 24.6250

More information

MACFP: Maximal Approximate Consecutive Frequent Pattern Mining under Edit Distance

MACFP: Maximal Approximate Consecutive Frequent Pattern Mining under Edit Distance MACFP: Maximal Approximate Consecutive Frequent Pattern Mining under Edit Distance Jingbo Shang, Jian Peng, Jiawei Han University of Illinois, Urbana-Champaign May 6, 2016 Presented by Jingbo Shang 2 Outline

More information

Branch Prediction based attacks using Hardware performance Counters IIT Kharagpur

Branch Prediction based attacks using Hardware performance Counters IIT Kharagpur Branch Prediction based attacks using Hardware performance Counters IIT Kharagpur March 19, 2018 Modular Exponentiation Public key Cryptography March 19, 2018 Branch Prediction Attacks 2 / 54 Modular Exponentiation

More information

Intractable Problems Part Two

Intractable Problems Part Two Intractable Problems Part Two Announcements Problem Set Five graded; will be returned at the end of lecture. Extra office hours today after lecture from 4PM 6PM in Clark S250. Reminder: Final project goes

More information

Distributed Box-Constrained Quadratic Optimization for Dual Linear SVM

Distributed Box-Constrained Quadratic Optimization for Dual Linear SVM Distributed Box-Constrained Quadratic Optimization for Dual Linear SVM Lee, Ching-pei University of Illinois at Urbana-Champaign Joint work with Dan Roth ICML 2015 Outline Introduction Algorithm Experiments

More information

TYPES OF MODEL COMPRESSION. Soham Saha, MS by Research in CSE, CVIT, IIIT Hyderabad

TYPES OF MODEL COMPRESSION. Soham Saha, MS by Research in CSE, CVIT, IIIT Hyderabad TYPES OF MODEL COMPRESSION Soham Saha, MS by Research in CSE, CVIT, IIIT Hyderabad 1. Pruning 2. Quantization 3. Architectural Modifications PRUNING WHY PRUNING? Deep Neural Networks have redundant parameters.

More information

Computing Correlation Anomaly Scores using Stochastic Nearest Neighbors

Computing Correlation Anomaly Scores using Stochastic Nearest Neighbors Computing Correlation Anomaly Scores using Stochastic Nearest Neighbors Tsuyoshi Idé IBM Research, Tokyo Research Laboratory Yamato, Kanagawa, Japan goodidea@jp.ibm.com Spiros Papadimitriou Michail Vlachos

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Saniv Kumar, Google Research, NY EECS-6898, Columbia University - Fall, 010 Saniv Kumar 9/13/010 EECS6898 Large Scale Machine Learning 1 Curse of Dimensionality Gaussian Mixture Models

More information

When Dictionary Learning Meets Classification

When Dictionary Learning Meets Classification When Dictionary Learning Meets Classification Bufford, Teresa 1 Chen, Yuxin 2 Horning, Mitchell 3 Shee, Liberty 1 Mentor: Professor Yohann Tendero 1 UCLA 2 Dalhousie University 3 Harvey Mudd College August

More information

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013. The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment 1 Caramanis/Sanghavi Due: Thursday, Feb. 7, 2013. (Problems 1 and

More information

Data Mining and Matrices

Data Mining and Matrices Data Mining and Matrices 08 Boolean Matrix Factorization Rainer Gemulla, Pauli Miettinen June 13, 2013 Outline 1 Warm-Up 2 What is BMF 3 BMF vs. other three-letter abbreviations 4 Binary matrices, tiles,

More information

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence Spring Announcements CS 188: Artificial Intelligence Spring 2010 Lecture 22: Nearest Neighbors, Kernels 4/18/2011 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements On-going: contest (optional and FUN!)

More information

Summary: A Random Walks View of Spectral Segmentation, by Marina Meila (University of Washington) and Jianbo Shi (Carnegie Mellon University)

Summary: A Random Walks View of Spectral Segmentation, by Marina Meila (University of Washington) and Jianbo Shi (Carnegie Mellon University) Summary: A Random Walks View of Spectral Segmentation, by Marina Meila (University of Washington) and Jianbo Shi (Carnegie Mellon University) The authors explain how the NCut algorithm for graph bisection

More information

Classification Semi-supervised learning based on network. Speakers: Hanwen Wang, Xinxin Huang, and Zeyu Li CS Winter

Classification Semi-supervised learning based on network. Speakers: Hanwen Wang, Xinxin Huang, and Zeyu Li CS Winter Classification Semi-supervised learning based on network Speakers: Hanwen Wang, Xinxin Huang, and Zeyu Li CS 249-2 2017 Winter Semi-Supervised Learning Using Gaussian Fields and Harmonic Functions Xiaojin

More information

Correlation Autoencoder Hashing for Supervised Cross-Modal Search

Correlation Autoencoder Hashing for Supervised Cross-Modal Search Correlation Autoencoder Hashing for Supervised Cross-Modal Search Yue Cao, Mingsheng Long, Jianmin Wang, and Han Zhu School of Software Tsinghua University The Annual ACM International Conference on Multimedia

More information

Logic BIST. Sungho Kang Yonsei University

Logic BIST. Sungho Kang Yonsei University Logic BIST Sungho Kang Yonsei University Outline Introduction Basics Issues Weighted Random Pattern Generation BIST Architectures Deterministic BIST Conclusion 2 Built In Self Test Test/ Normal Input Pattern

More information

High Dimensional Search Min- Hashing Locality Sensi6ve Hashing

High Dimensional Search Min- Hashing Locality Sensi6ve Hashing High Dimensional Search Min- Hashing Locality Sensi6ve Hashing Debapriyo Majumdar Data Mining Fall 2014 Indian Statistical Institute Kolkata September 8 and 11, 2014 High Support Rules vs Correla6on of

More information

CMPS 6630: Introduction to Computational Biology and Bioinformatics. Structure Comparison

CMPS 6630: Introduction to Computational Biology and Bioinformatics. Structure Comparison CMPS 6630: Introduction to Computational Biology and Bioinformatics Structure Comparison Protein Structure Comparison Motivation Understand sequence and structure variability Understand Domain architecture

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu 10/24/2012 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu

More information

A New Space for Comparing Graphs

A New Space for Comparing Graphs A New Space for Comparing Graphs Anshumali Shrivastava and Ping Li Cornell University and Rutgers University August 18th 2014 Anshumali Shrivastava and Ping Li ASONAM 2014 August 18th 2014 1 / 38 Main

More information

Design and Implementation of Speech Recognition Systems

Design and Implementation of Speech Recognition Systems Design and Implementation of Speech Recognition Systems Spring 2013 Class 7: Templates to HMMs 13 Feb 2013 1 Recap Thus far, we have looked at dynamic programming for string matching, And derived DTW from

More information

CS145: INTRODUCTION TO DATA MINING

CS145: INTRODUCTION TO DATA MINING CS145: INTRODUCTION TO DATA MINING 5: Vector Data: Support Vector Machine Instructor: Yizhou Sun yzsun@cs.ucla.edu October 18, 2017 Homework 1 Announcements Due end of the day of this Thursday (11:59pm)

More information

ECS289: Scalable Machine Learning

ECS289: Scalable Machine Learning ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Oct 27, 2015 Outline One versus all/one versus one Ranking loss for multiclass/multilabel classification Scaling to millions of labels Multiclass

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

Caesar s Taxi Prediction Services

Caesar s Taxi Prediction Services 1 Caesar s Taxi Prediction Services Predicting NYC Taxi Fares, Trip Distance, and Activity Paul Jolly, Boxiao Pan, Varun Nambiar Abstract In this paper, we propose three models each predicting either taxi

More information

CS345a: Data Mining Jure Leskovec and Anand Rajaraman Stanford University

CS345a: Data Mining Jure Leskovec and Anand Rajaraman Stanford University CS345a: Data Mining Jure Leskovec and Anand Rajaraman Stanford University TheFind.com Large set of products (~6GB compressed) For each product A=ributes Related products Craigslist About 3 weeks of data

More information

Matrices, Vector Spaces, and Information Retrieval

Matrices, Vector Spaces, and Information Retrieval Matrices, Vector Spaces, and Information Authors: M. W. Berry and Z. Drmac and E. R. Jessup SIAM 1999: Society for Industrial and Applied Mathematics Speaker: Mattia Parigiani 1 Introduction Large volumes

More information

RaRE: Social Rank Regulated Large-scale Network Embedding

RaRE: Social Rank Regulated Large-scale Network Embedding RaRE: Social Rank Regulated Large-scale Network Embedding Authors: Yupeng Gu 1, Yizhou Sun 1, Yanen Li 2, Yang Yang 3 04/26/2018 The Web Conference, 2018 1 University of California, Los Angeles 2 Snapchat

More information

Slides based on those in:

Slides based on those in: Spyros Kontogiannis & Christos Zaroliagis Slides based on those in: http://www.mmds.org High dim. data Graph data Infinite data Machine learning Apps Locality sensitive hashing PageRank, SimRank Filtering

More information

Nearest Neighbor Search with Keywords in Spatial Databases

Nearest Neighbor Search with Keywords in Spatial Databases 776 Nearest Neighbor Search with Keywords in Spatial Databases 1 Sphurti S. Sao, 2 Dr. Rahila Sheikh 1 M. Tech Student IV Sem, Dept of CSE, RCERT Chandrapur, MH, India 2 Head of Department, Dept of CSE,

More information

Probabilistic Similarity Search for Uncertain Time Series

Probabilistic Similarity Search for Uncertain Time Series Probabilistic Similarity Search for Uncertain Time Series Johannes Aßfalg, Hans-Peter Kriegel, Peer Kröger, Matthias Renz Ludwig-Maximilians-Universität München, Oettingenstr. 67, 80538 Munich, Germany

More information

Nonsmooth Analysis and Subgradient Methods for Averaging in Dynamic Time Warping Spaces

Nonsmooth Analysis and Subgradient Methods for Averaging in Dynamic Time Warping Spaces Nonsmooth Analysis and Subgradient Methods for Averaging in Dynamic Time Warping Spaces David Schultz and Brijnesh Jain Technische Universität Berlin, Germany arxiv:1701.06393v1 [cs.cv] 23 Jan 2017 Abstract.

More information

Phylogenetic Tree Reconstruction

Phylogenetic Tree Reconstruction I519 Introduction to Bioinformatics, 2011 Phylogenetic Tree Reconstruction Yuzhen Ye (yye@indiana.edu) School of Informatics & Computing, IUB Evolution theory Speciation Evolution of new organisms is driven

More information