Time Series Shapelets

Similar documents
Learning Time-Series Shapelets

Recent advances in Time Series Classification

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

Time Series Classification

Data Mining Prof. Pabitra Mitra Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur

Homework 1 Solutions Probability, Maximum Likelihood Estimation (MLE), Bayes Rule, knn

Linear and Logistic Regression. Dr. Xiaowei Huang

Landslide Classification: An Object-Based Approach Bryan Zhou Geog 342: Final Project

Will it rain tomorrow?

Machine Learning on temporal data

Final Exam, Machine Learning, Spring 2009

Data Mining. 3.6 Regression Analysis. Fall Instructor: Dr. Masoud Yaghini. Numeric Prediction

LEC1: Instance-based classifiers

Mining Classification Knowledge

Symbolic methods in TC: Decision Trees

Mining Classification Knowledge

Jeffrey D. Ullman Stanford University

Predictive Modeling: Classification. KSE 521 Topic 6 Mun Yi

The Solution to Assignment 6

Midterm Review CS 7301: Advanced Machine Learning. Vibhav Gogate The University of Texas at Dallas

Holdout and Cross-Validation Methods Overfitting Avoidance

CLUe Training An Introduction to Machine Learning in R with an example from handwritten digit recognition

Decision Trees. Introduction. Some facts about decision trees: They represent data-classification models.

Decision Tree Learning Mitchell, Chapter 3. CptS 570 Machine Learning School of EECS Washington State University

Search for the Dubai in the remap search bar:

Data Mining and Knowledge Discovery: Practice Notes

Introduction to Machine Learning Midterm Exam

HOT SAX: Efficiently Finding the Most Unusual Time Series Subsequence

Skylines. Yufei Tao. ITEE University of Queensland. INFS4205/7205, Uni of Queensland

10701/15781 Machine Learning, Spring 2007: Homework 2

Unit 5 Evaluation. Multiple-Choice. Evaluation 05 Second Year Algebra 1 (MTHH ) Name I.D. Number

Classification: Decision Trees

Review of Lecture 1. Across records. Within records. Classification, Clustering, Outlier detection. Associations

Machine Learning Summer 2018 Exercise Sheet 4

Decision Trees. CS57300 Data Mining Fall Instructor: Bruno Ribeiro

Data Mining. CS57300 Purdue University. Bruno Ribeiro. February 8, 2018

Machine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20.

Search for the Dubai in the remap search bar:

EECS 349:Machine Learning Bryan Pardo

CS145: INTRODUCTION TO DATA MINING

Nearest Neighbor Search with Keywords

ECE 661: Homework 10 Fall 2014

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION

Assignment 1: Probabilistic Reasoning, Maximum Likelihood, Classification

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning

Information Extraction from Text

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

Decision Tree Learning

Generalization Error on Pruning Decision Trees

6.034 Quiz 3 November 12, Thu. Time Instructor. 1-2 Bob Berwick. 2-3 Bob Berwick. 3-4 Bob Berwick

CSE 546 Final Exam, Autumn 2013

Weather Prediction Using Historical Data

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas

FINAL: CS 6375 (Machine Learning) Fall 2014

Principles of Pattern Recognition. C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata

Quickest detection & classification of transient stability status for power systems

Online Advertising is Big Business

COMS 4771 Introduction to Machine Learning. Nakul Verma

Midterm: CS 6375 Spring 2015 Solutions

Machine Learning Basics

Style-aware Mid-level Representation for Discovering Visual Connections in Space and Time

Rule Generation using Decision Trees

Classification with Perceptrons. Reading:

Midterm: CS 6375 Spring 2018

Decision Support Systems MEIC - Alameda 2010/2011. Homework #8. Due date: 5.Dec.2011

ELEC6910Q Analytics and Systems for Social Media and Big Data Applications Lecture 3 Centrality, Similarity, and Strength Ties

The Deep Forest and its Modi cations

Geometric View of Machine Learning Nearest Neighbor Classification. Slides adapted from Prof. Carpuat

FINAL EXAM: FALL 2013 CS 6375 INSTRUCTOR: VIBHAV GOGATE

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS

Generative Models. CS4780/5780 Machine Learning Fall Thorsten Joachims Cornell University

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

Nonlinear Classification

From perceptrons to word embeddings. Simon Šuster University of Groningen

ECE 592 Topics in Data Science

Name (NetID): (1 Point)

CSCE 478/878 Lecture 6: Bayesian Learning

CSCI-567: Machine Learning (Spring 2019)

Università di Pisa A.A Data Mining II June 13th, < {A} {B,F} {E} {A,B} {A,C,D} {F} {B,E} {C,D} > t=0 t=1 t=2 t=3 t=4 t=5 t=6 t=7

Name I.D. Number. Select the response that best completes the statement or answers the question.

Insect ID. Antennae Length. Insect Class. Abdomen Length

Let's look at some higher order equations (cubic and quartic) that can also be solved by factoring.

Midterm. You may use a calculator, but not any device that can access the Internet or store large amounts of data.

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Decision Trees. Tobias Scheffer

Analysis of Data Mining Techniques for Weather Prediction

Deep Learning. Convolutional Neural Networks Applications

1 [15 points] Frequent Itemsets Generation With Map-Reduce

Decision Trees Part 1. Rao Vemuri University of California, Davis

Notes on Machine Learning for and

CS246 Final Exam. March 16, :30AM - 11:30AM

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II)

Data Mining Classification: Basic Concepts and Techniques. Lecture Notes for Chapter 3. Introduction to Data Mining, 2nd Edition

Machine Learning Lecture 7

Equations and Inequalities

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang

Machine Learning for NLP

MIDTERM SOLUTIONS: FALL 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE

Online Videos FERPA. Sign waiver or sit on the sides or in the back. Off camera question time before and after lecture. Questions?

Uncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization

Online Courses for High School Students

Transcription:

Time Series Shapelets Lexiang Ye and Eamonn Keogh This file contains augmented versions of the figures in our paper, plus additional experiments and details that were omitted due to space limitations Please recall that you can visit our webpage for even more information, including all code and all datasets. www.cs.ucr.edu/~lexiangy/shapelet.html Until the paper is accepted, the webpage is password-protected with the username/password:kdd / riverside Recall that there are many tools to allow anonymous surfing, we suggest you Google anonymous surfing

Scalability Test Experiment: Part 1 5 *1 5 About 5 days.98 seconds 4 *1 5 3 *1 5 2 *1 5 1 *1 5 Brute Force Early Abandon Pruning Entropy Pruning Combined Pruning accuracy.96.94.92 Currently best published accuracy 91.1% 1 2 4 8 16.9 1 2 4 8 16 The dataset came from Jeffery, C. (25). Synthetic Lightning EMP Data. http://public.lanl.gov/eads/datasets/emp/index.html Los Alamos National Laboratory. This dataset is also mentioned in http://www.cs.ucr.edu/~eamonn/isax/isax_weighted.ppt Figure in Paper Raw Numbers Size Full Speedup Distance Early Abandon Entropy Prune Brute Force Accuracy seconds 5 *1 5 4 *1 5 3 *1 5 2 *1 5 1 *1 5 About 5 days Brute Force Early Abandon Pruning Entropy Pruning Combined Pruning accuracy.98.96.94.92 Currently best published accuracy 91.1% 1 499 118 739 1664.936 2 162 4524 1571 6662.936 4 2236 18362 3241 26736.9688 1 2 4 8 16.9 1 2 4 8 16 8 4356 7436 6324 17.958 16 8868 2449 12932 429.9718

Scalability Test Experiment: Part 2 Here is the decision tree learned on this problem Shapelet Dictionary 97.8 1 2 3.3.2.1 From the shapelet, we can see the most distinguishing feature for time series objects from two classes is: the lightning in class A has a shallow slope followed by a steep one; the lightning part in class B has a directly steep slope. Lightning Decision Tree 1 1.2.8.4 1.2.6 Class A Class B -.4 4 8 12 16 2 -.6 4 8 12 16 2 Shapelet in the time series object Comparison of time series objects from two classes

Projectile Points Experiment: Part 1 seconds 9 6 3 Brute Force Early Abandon Pruning Entropy Pruning Combined Pruning accuracy.8.7.6 accuracy of rotation invariant nearest neighbor classifier is 68% 3 6 12 36.5 3 6 12 36 The dataset can be downloaded from the website: http://www.cs.ucr.edu/~lexiangy/shapelet.html Raw Numbers Size Combined Pruning Distance Early Abandon Entropy Prune Brute Force Accuracy 3 29.81 29.95 49.28 49.51.5257 6 114.78 118.85 184.86 193.75.64 12 352 486.78 541.37 779.54.6828 36 1722 51.94 259.74 778.42.8

Projectile Points Experiment: Part 2 Here is the decision tree learned on this problem Clovis Avonlea the Clovis projectile points can be distinguished from the others by an un-notched hafting area near the bottom connected by a deep concave bottom end. After distinguishing the Clovis projectile points, the Avonlea points are differentiated from the mixed class by a small notched hafting area connected by a shallow concave bottom end. (Clovis) 11.24 1.5 1..5 (Avonlea) 85.47 Shapelet Dictionary 1 2 3 4 Arrowhead Decision Tree trainingc 36 test 175 length 45* shapelet length ndex start pos 194 35 123 1 2 221 25 184 *: the length of the arrowheads varies, since it is not normalization by lengths. **: the corresponding shapelet can be gotten using the index and start pos in the training dataset. ndex starts from 1.

Projectile Points Experiment: Part 3 Method Break Ties Specify Dist. Measure Accuracy Longest.749 Classification time Shapelet Shortest.469 Max Separation.721 Max Mean Separation.8.332 sec. Nearest Neighbor Rotation invariant Euclidean Dist. 68 113 sec. DTW

Projectile Points Experiment: Part 4 Here are the projectile points that the shapelet decision tree classifier classifies correctly and the rotation-invariant nearest neighbor classifier does not.

Projectile Points Experiment mage process Adjust the images to approximately same size Convert to time series using the angle-based method Concatenate a copy of the time series to itself. n this way, we make it rotation invariant fixed angle size θ Why not normalize the length? Normalized to the same length makes the similar part of outline of different length very different unmatched.2.4.6.8 1.2.4.6.8 1

Projectile Points Reference n the paper we considered the problem of projectile point (arrowhead) classification. Space limitations prevented an extensive discussion of the topic, a detailed review of the literature on projectile point classification can be downloaded from a link on our webpage. More detailed information is from http://www.museum.state.il.us/ismdepts/anthro/proj_point/points_glossary.html

Mining Historical Documents Experiment: Part 1 seconds 1.8 1 5 1.2 1 5 6 1 4 Brute Force Early Abandon Pruning Entropy Pruning Combined Pruning accuracy 1.9.8.7.6 accuracy of rotation invariant nearest neighbor classifier is 89.2% 3 6 12 3.5 3 6 12 3 The dataset can be downloaded from the website: http://www.cs.ucr.edu/~lexiangy/shapelet.html Raw Numbers Size Combined Pruning Distance Early Abandon Entropy Prune Brute Force Accuracy 3 856.84 857.66 1554.69 1551.42.5271 6 3653.53 3793.27 658.92 6799.2.539 12 12292.2 15697.6 21944.8 27737.8217 3 54359.6 95654.17 97823.7 17512.5.8992

Mining Historical Documents Experiment: Part 2 Here is the decision tree learned on this problem (Polish) (Spanish) 151.7 156.1 Polish Spanish 3 2 1 From the shapelets, we can see the most distinguishing features for time series objects for three classes is: The Polish examples are distinguished by a curved edge which has a sharp angle at the very bottom; the unique semi-circular bottom of the Spanish crest is used in node to discriminate it from the French examples. Shaplet Dictionary 1 2 3 4 Shield Decision Tree trainingc 3 test 129 length 11 1 2 Shapelet** length ndex start pos 164 11 267 234 12 275 training / testing: 3 / 129 *: the length of the shields varies, since it is not normalization by lengths **: the corresponding shapelet can be gotten using the index and start pos in the training dataset. ndex starts from 1.

Mining Historical Documents Experiment: Part 3 Method Break Ties Specify Dist. Measure Accuracy Longest.767 Classification time Shapelet Shortest.698 Max Separation.86 Max Mean Separation.899.48 sec. Nearest Neighbor Rotation invariant Euclidean Dist.829 156 sec. DTW

Mining Historical Documents Experiment: Part 4 Here is the shields that the shapelet decision tree classifier classifies correctly and the rotation-invariant nearest neighbor classifier does not. Note it can handle broken shapes.

Historical Documents Reference Here is the full version of the cropped page from [a] which appears in the paper [a] Montagu, J.A. (184). A guide to the study of heraldry. Publisher: London : W. Pickering. Online version www.archive.org/details/guidetostudyofhe montuoft

Gun / NoGun Experiment: Part 1 seconds 25 2 15 1 5 Brute Force Early Abandon Pruning Entropy Pruning Combined Pruning accuracy.8.7.6 accuracy of rotation invariant nearest neighbor classifier is 68% 6 12 24 5.5 6 12 24 5 The dataset can be downloaded from the website: http://www.cs.ucr.edu/~lexiangy/shapelet.html Raw Numbers Size Combined Pruning Distance Early Abandon Entropy Prune Brute Force Accuracy 3 29.81 29.95 49.28 49.51.5257 6 114.78 118.85 184.86 193.75.64 12 352 486.78 541.37 779.54.6828 36 1722 51.94 259.74 778.42.8

Gun / NoGun Experiment: Part 2 Here is the decision tree learned on this problem No Gun Gun the NoGun class has a dip where the actor put her hand down by her side, and inertia carries her hand a little too far and she is forced to correct for it (a phenomenon known as overshoot ). n contrast, when the actor has the gun, she returns her hand to her side more carefully, feeling for the gun holster, and no dip is seen. (No Gun) 38.94 3 2 1 Shaplet Dictionary 25 5 75 1 Gun Decision Tree trainingc 5 test 15 1 length 15 Shapelet* length ndex start pos 38 45 19 training / testing: 5 / 15 *: the corresponding shapelet can be gotten using the index and start pos in the training dataset. ndex starts from 1.

Gun / NoGun Experiment: Part 3 Method Break Ties Specify Dist. Measure Accuracy Classification time Longest.933.16 sec. Shapelet Shortest.69 Max Separation.933 Max Mean Separation.893 Euclidean Dist.913.64 sec. Nearest Neighbor DTW best warping window DTW no warping window.913.97

Wheat Spectrography Experiment: Part 1 seconds 6 4 2 Brute Force Early Abandon Pruning Entropy Pruning Combined Pruning 14 28 49 accuracy.8.7.6.5.4 accuracy of rotation invariant nearest neighbor classifier is 54.3% 14 28 49 The dataset can be downloaded from the website: http://www.cs.ucr.edu/~lexiangy/shapelet.html Raw Numbers Size Combined Pruning Distance Early Abandon Entropy Prune Brute Force Accuracy 14 21.96 22326.3 35678.2 37345.1.497 28 55716.31 92.7 132786.8 149992.1.533 49 122456 264211.5 343211.7 789533.72

Wheat Spectrography Experiment: Part 2 Here is the decision tree learned on this problem Shapelet Dictionary.4.3 5 6 4 3 2 1 1.5.2 2 4 6 8 1 12 V V V.1. 1 2 3 Wheat Decision Tree sample of time series object from each class trainingc 49 test 276 length 15 Shapelet length ndex start pos V 1 21 7 V V 8 47 764 2 4 1 3 6 5 5 33 958 V 5 8 958 V 2 7 59 V 26 42 496

Wheat Spectrography Experiment: Part 3 Method Break Ties Specify Dist. Measure Accuracy Longest.53 Classification time Shapelet Shortest.547 Max Separation.72 1.428 sec. Max Mean Separation.77 Nearest Neighbor Euclidean Dist.543.36 sec.

MALLET Experiment: Part 1 Here is the decision tree learned on this problem Shapelet Dictionary data sample from each class* 39.3 2.42 34.9 4.3 2.9 51.6.2 V V V V 6 4 2 5 1 15 2 25 MALLET Decision Tree V V V 2 4 5 6 1 7 3 The first shapelet covers the first rectangular dip. All the data objects on the left has either no or shallow dip, while objects on the right of the tree has deeper dip. Similar on the third shapelet The fourth shapelet shows that the most critical difference between the objects in class 5 and 6 is whether the second and the third peaks are present. Similar difference class 1 and 7 are interpreted by the sixth shapelet. The fifth shapelet presents the difference in the last peak. The objects on the left have a deep dip and those on right have no or shallow dip. V * The figures are in the reverse order of time series data. trainingc 8 test 232 length 256 Shapelet length ndex start pos 49 75 27 15 13 165 5 27 25 V 84 14 115 V 36 58 194 V 87 18 111 V 21 28 74

MALLET Experiment: Part 2 Method Break Ties Specify Dist. Measure Accuracy Longest.99 Classification time Shapelet Shortest.74 Max Separation.945 Max Mean Separation.947 Euclidean Dist.959 Nearest Neighbor DTW best warping window DTW no warping window.963.951