Anomaly Detection for the CERN Large Hadron Collider injection magnets

Size: px
Start display at page:

Download "Anomaly Detection for the CERN Large Hadron Collider injection magnets"

Transcription

1 Anomaly Detection for the CERN Large Hadron Collider injection magnets Armin Halilovic KU Leuven - Department of Computer Science In cooperation with CERN

2 0 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 1 Anomaly Detection for the CERN Large Hadron Collider injection magnets

3 1 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 2 Anomaly Detection for the CERN Large Hadron Collider injection magnets

4 1 Context - Anomaly Detection Classification Normal vs. Abnormal/novel data One-class vs. Multiclass classification High amount of normal data Very low amount of anomalous data Unsupervised machine learning models Assign anomaly scores to data Outlier removal 3 Anomaly Detection for the CERN Large Hadron Collider injection magnets

5 1 Context - Problem Statement & Motivation The goal is to develop an anomaly detection application that can detect anomalies in the behaviour of the injection kicker magnets of the Large Hadron Collider. This is useful, because it can be used to: Detect anomalous behaviour and thus predict failures Improve CERN s response time Improve machine reliability 4 Anomaly Detection for the CERN Large Hadron Collider injection magnets

6 2 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 5 Anomaly Detection for the CERN Large Hadron Collider injection magnets

7 2 Data - Types I 6 types of data collections: 1 Continuous 2 Internal Post Operational Check (IPOC) 3 State 4 Controller 5 LHC 6 Electronic Logbook Continuous & discrete variables Fixed sampling rates & asynchronous sampling triggers 120 data collections Data from June 2015 to September Anomaly Detection for the CERN Large Hadron Collider injection magnets

8 2 Data - Types II Continuous Data: Temperature and pressures Fixed frequency sampling + save based on change in value Missing data: Forward Fill 7 Anomaly Detection for the CERN Large Hadron Collider injection magnets

9 2 Data - Types III Continuous Data: Temperature and pressures Fixed frequency sampling + save based on change in value Missing data: Forward Fill 8 Anomaly Detection for the CERN Large Hadron Collider injection magnets

10 2 Data - Types IV Internal Post Operational Check (IPOC) Data: Closely related to magnets: energy, strength, delay,... Only sampled when magnet generators pulse All IPOC measurements recorded simultaneously At most once every 10 seconds Many large gaps when experiments run Missing data: cannot fill Different timestamps for beams B1 and B2 Anomaly detection for the two MKI installations separately 9 Anomaly Detection for the CERN Large Hadron Collider injection magnets

11 2 Data - Types V IPOC, I STRENGTH, 2016: 10 Anomaly Detection for the CERN Large Hadron Collider injection magnets

12 2 Data - Types VI State Data: Not used No data for 2015 Controller Data: Not used Technical issues (duplicate timestamps) with received database 11 Anomaly Detection for the CERN Large Hadron Collider injection magnets

13 2 Data - Types VII LHC Data: Particle beam measurements: beam intensity & beam length Sampled and stored in similar way to Continuous measurements Missing data: Forward fill 12 Anomaly Detection for the CERN Large Hadron Collider injection magnets

14 2 Data - Types VIII Electronic Logbook Data: Manually created logbook entries (labels) Describe certain events Anomaly labels not precise, but range of 12 hours Label type Beam 1 Beam 2 anomaly fault info intervention research Total: Anomaly Detection for the CERN Large Hadron Collider injection magnets

15 2 Data - IPOC Segments I Magnets only in use for certain time periods IPOC data sampled only when magnets in use IPOC segment = period of magnet usage Introduced to deal with uncertainty of anomaly labels Important semantic meaning Data is split into segments based on segmentation distance 14 Anomaly Detection for the CERN Large Hadron Collider injection magnets

16 2 Data - IPOC Segments II Data is split into segments based on segmentation distance 15 Anomaly Detection for the CERN Large Hadron Collider injection magnets

17 3 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 16 Anomaly Detection for the CERN Large Hadron Collider injection magnets

18 3 Preprocessing - Data Filtering I Want to train models based on correct/relevant data Sudden extremely high temperatures, negative timing, etc. are impossible Measurement Minimum Maximum PRESSURE mbar mbar TEMP MAGNET (DOWN UP) 18 C 60 C TEMP TUBE (DOWN UP) 18 C 120 C I STRENGTH 1 ka N/A T DELAY 10 µs N/A 17 Anomaly Detection for the CERN Large Hadron Collider injection magnets

19 3 Preprocessing - Data Filtering II True pattern emerges 18 Anomaly Detection for the CERN Large Hadron Collider injection magnets

20 3 Preprocessing - Data Filtering III Impossible time delays removed 19 Anomaly Detection for the CERN Large Hadron Collider injection magnets

21 3 Preprocessing - Features All IPOC data + Continuous data at IPOC data timestamps (with forward fill) + LHC data at IPOC data timestamps (with forward fill) + Temporal features on Continuous and LHC data: To catch temporal relationship in data Sliding window features: mean & sum Important parameter: sliding window size Done separately for both B1 and B2 20 Anomaly Detection for the CERN Large Hadron Collider injection magnets

22 4 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 21 Anomaly Detection for the CERN Large Hadron Collider injection magnets

23 4 Anomaly Detection Train machine learning model using preprocessed data Use the model to generate anomaly scores Rescale scores to [0, 1] 22 Anomaly Detection for the CERN Large Hadron Collider injection magnets

24 4 Anomaly Detection - Isolation Forest Anomaly Scores 23 Anomaly Detection for the CERN Large Hadron Collider injection magnets

25 4 Anomaly Detection - Gaussian Mixture Model Scores I 24 Anomaly Detection for the CERN Large Hadron Collider injection magnets

26 4 Anomaly Detection - Gaussian Mixture Model Scores II 25 Anomaly Detection for the CERN Large Hadron Collider injection magnets

27 4 Anomaly Detection - Dummy Detectors Simple detection strategies as baseline to compare to Constant, uniformly random, stratified random 26 Anomaly Detection for the CERN Large Hadron Collider injection magnets

28 5 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 27 Anomaly Detection for the CERN Large Hadron Collider injection magnets

29 5 Postprocessing I Anomaly labels are unspecific, 12 hour range Will use segments instead of individual data tuples in evaluation Transform scored data into lists of IPOC segments Segment anomaly score based on anomaly scores of its data Anomalous behavior likely occurs in multiple successive timestamps These timestamps should get higher anomaly scores The segments that contain these timestamps should then have higher anomaly scores 28 Anomaly Detection for the CERN Large Hadron Collider injection magnets

30 5 Postprocessing II Methods for Segment Anomaly Score: Max Top K (10) Top Percentage (25%) Ground Truth Annotation: Need to compare segment anomaly scores to consistent basis of ground truth This allows for fair performance evaluation Mark segments as anomalous if they lie in the 12 hour range of an anomaly label 29 Anomaly Detection for the CERN Large Hadron Collider injection magnets

31 5 Postprocessing III We now have: A set of IPOC segments with anomaly scores Knowledge of which segments are actually anomalous 30 Anomaly Detection for the CERN Large Hadron Collider injection magnets

32 6 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 31 Anomaly Detection for the CERN Large Hadron Collider injection magnets

33 6 Evaluation Anomaly scores lie in [0, 1] Ground truth is 0 or 1 To evaluate performance, need to select a threshold anomaly score in order to count True Positives, False Positives, True Negatives, and False Negatives If score above threshold, then prediction is Positive, else Negative Prediction Ground Truth Positive Negative Positive TP FP Negative FN TN 32 Anomaly Detection for the CERN Large Hadron Collider injection magnets

34 6 Evaluation - Performance Metric Precision and Recall are useful context of imbalanced data P recision = T P T P +F P Recall = T P T P +F N But, want single number as performance metric for automated comparisons Calculate Precision and Recall for each possible anomaly score threshold and plot the resulting curve Performance metric = Area under Precision-Recall Curve (AUPR) 33 Anomaly Detection for the CERN Large Hadron Collider injection magnets

35 6 Evaluation - Grid Search Many parameters for developed anomaly detection pipeline Segmentation distance, scale data, anomaly score method, anomaly detector, anomaly detector hyperparameters, labels Grid search for parameter optimization Pipeline is executed automatically with predetermined combinations of parameters built by a grid of parameters Results are stored and sorted by AUPR so that the best performing parameters can be found easily 34 Anomaly Detection for the CERN Large Hadron Collider injection magnets

36 7 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 35 Anomaly Detection for the CERN Large Hadron Collider injection magnets

37 7 Results - Dummy Detectors Figure: PR curves of Dummy detectors with evaluation parameters segmentation distance = 30 min, anomaly score method = top k, labels = all. 36 Anomaly Detection for the CERN Large Hadron Collider injection magnets

38 7 Results - GMM I Best PR Curve Figure: Parameters: n components = 6, covariance type = full, scale data = F alse, segmentation distance = 60 min, anomaly score method = top k, labels = anomaly 37 Anomaly Detection for the CERN Large Hadron Collider injection magnets

39 7 Results - GMM II Predictions with 99-th percentile anomaly score threshold 38 Anomaly Detection for the CERN Large Hadron Collider injection magnets

40 7 Results - Isolation Forest I Best PR Curve Figure: Parameters: n estimators = 250, max samples = 5120, scale data = F alse, segmentation distance = 60 min, anomaly score method = max, labels = anomaly 39 Anomaly Detection for the CERN Large Hadron Collider injection magnets

41 7 Results - Isolation Forest II Predictions with 99-th percentile anomaly score threshold 40 Anomaly Detection for the CERN Large Hadron Collider injection magnets

42 7 Results - Isolation Forest III Only IPOC features, best PR Curve 41 Anomaly Detection for the CERN Large Hadron Collider injection magnets

43 7 Results - Isolation Forest IV Only IPOC features, predictions with 99-th percentile anomaly score threshold 42 Anomaly Detection for the CERN Large Hadron Collider injection magnets

44 7 Results - Source Code Written to be extensible Pipeline components in clear modules: preprocessing anomaly detection postprocessing evaluation pipeline Parameters can be varied easily 43 Anomaly Detection for the CERN Large Hadron Collider injection magnets

45 8 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 44 Anomaly Detection for the CERN Large Hadron Collider injection magnets

46 8 Conclusion Anomaly detection application has been developed Some anomalies are detected very well Many are still not detected at all Experiments have shown that performance can still be improved significantly More experiments should be done around feature selection 45 Anomaly Detection for the CERN Large Hadron Collider injection magnets

47 8 Future Work Feature selection Controller data Integration of more anomaly detectors (e.g. one class SVM or Local Outlier Factor) Better segmentation procedure without segmentation distance parameter More efficient and autonomous parameter optimization using e.g. Evolutionary algorithms or Bayesian Optimization 46 Anomaly Detection for the CERN Large Hadron Collider injection magnets

48 8 Bibliography CERN. Overview lhc. Accessed W Herr and T Pieloni. Beam-beam effects. (arxiv: ):1 29, Contribution to the CAS - CERN Accelerator School: Advanced Accelerator Physics Course, Trondheim, Norway, Aug Fei Tony Liu, Kai Ming Ting, and Zhi-Hua Zhou. Isolation forest. In Data Mining, ICDM 08. Eighth IEEE International Conference on, pages IEEE, Anomaly Detection for the CERN Large Hadron Collider injection magnets

49 Questions?

50 8 Extra - Comparison to Previous Work Enabled use of many machine learning models instead of just 1 Segmentation of input data instead of segmentation of output anomaly scores Consistent basis of ground truth more correct comparison of results Evaluation metrics in terms of TP, FP TN, FN instead of ambiguous terms PR curve using all anomaly score thresholds instead of calculating Precision and Recall for one threshold Anomaly Detection for the CERN Large Hadron Collider injection magnets

51 8 Extra - Isolation Forest Ensemble of simple decision trees which split randomly on features Trees are grown for random samples of dataset until each data tuple forms a leaf node Average path length will be shorter for anomalies Works well in high dimensional problems Density estimation, but without a density measure Source: [3] 49 Anomaly Detection for the CERN Large Hadron Collider injection magnets

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 1 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification

More information

Data Mining Based Anomaly Detection In PMU Measurements And Event Detection

Data Mining Based Anomaly Detection In PMU Measurements And Event Detection Data Mining Based Anomaly Detection In PMU Measurements And Event Detection P. Banerjee, S. Pandey, M. Zhou, A. Srivastava, Y. Wu Smart Grid Demonstration and Research Investigation Lab (SGDRIL) Energy

More information

Class 4: Classification. Quaid Morris February 11 th, 2011 ML4Bio

Class 4: Classification. Quaid Morris February 11 th, 2011 ML4Bio Class 4: Classification Quaid Morris February 11 th, 211 ML4Bio Overview Basic concepts in classification: overfitting, cross-validation, evaluation. Linear Discriminant Analysis and Quadratic Discriminant

More information

Machine Learning Concepts in Chemoinformatics

Machine Learning Concepts in Chemoinformatics Machine Learning Concepts in Chemoinformatics Martin Vogt B-IT Life Science Informatics Rheinische Friedrich-Wilhelms-Universität Bonn BigChem Winter School 2017 25. October Data Mining in Chemoinformatics

More information

CS145: INTRODUCTION TO DATA MINING

CS145: INTRODUCTION TO DATA MINING CS145: INTRODUCTION TO DATA MINING 4: Vector Data: Decision Tree Instructor: Yizhou Sun yzsun@cs.ucla.edu October 10, 2017 Methods to Learn Vector Data Set Data Sequence Data Text Data Classification Clustering

More information

Evaluation & Credibility Issues

Evaluation & Credibility Issues Evaluation & Credibility Issues What measure should we use? accuracy might not be enough. How reliable are the predicted results? How much should we believe in what was learned? Error on the training data

More information

The MD was done at 450GeV using beam 2 only. An MD focussing on injection of bunches with nominal emittance was done in parallel on beam 1.

The MD was done at 450GeV using beam 2 only. An MD focussing on injection of bunches with nominal emittance was done in parallel on beam 1. CERN-ATS-Note-2011-065 MD 2011-08-08 Tobias.Baer@cern.ch MKI UFOs at Injection Tobias BAER, Mike BARNES, Wolfgang BARTMANN, Chiara BRACCO, Etienne CARLIER, Christophe CHANAVAT, Lene Norderhaug DROSDAL,

More information

Evaluation. Andrea Passerini Machine Learning. Evaluation

Evaluation. Andrea Passerini Machine Learning. Evaluation Andrea Passerini passerini@disi.unitn.it Machine Learning Basic concepts requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain

More information

Anomaly Detection. Jing Gao. SUNY Buffalo

Anomaly Detection. Jing Gao. SUNY Buffalo Anomaly Detection Jing Gao SUNY Buffalo 1 Anomaly Detection Anomalies the set of objects are considerably dissimilar from the remainder of the data occur relatively infrequently when they do occur, their

More information

A New Unsupervised Event Detector for Non-Intrusive Load Monitoring

A New Unsupervised Event Detector for Non-Intrusive Load Monitoring A New Unsupervised Event Detector for Non-Intrusive Load Monitoring GlobalSIP 2015, 14th Dec. Benjamin Wild, Karim Said Barsim, and Bin Yang Institute of Signal Processing and System Theory of,, Germany

More information

Evaluation requires to define performance measures to be optimized

Evaluation requires to define performance measures to be optimized Evaluation Basic concepts Evaluation requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain (generalization error) approximation

More information

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology

More information

Unsupervised Anomaly Detection for High Dimensional Data

Unsupervised Anomaly Detection for High Dimensional Data Unsupervised Anomaly Detection for High Dimensional Data Department of Mathematics, Rowan University. July 19th, 2013 International Workshop in Sequential Methodologies (IWSM-2013) Outline of Talk Motivation

More information

Lecture 3: Decision Trees

Lecture 3: Decision Trees Lecture 3: Decision Trees Cognitive Systems - Machine Learning Part I: Basic Approaches of Concept Learning ID3, Information Gain, Overfitting, Pruning last change November 26, 2014 Ute Schmid (CogSys,

More information

15-388/688 - Practical Data Science: Nonlinear modeling, cross-validation, regularization, and evaluation

15-388/688 - Practical Data Science: Nonlinear modeling, cross-validation, regularization, and evaluation 15-388/688 - Practical Data Science: Nonlinear modeling, cross-validation, regularization, and evaluation J. Zico Kolter Carnegie Mellon University Fall 2016 1 Outline Example: return to peak demand prediction

More information

Midterm: CS 6375 Spring 2015 Solutions

Midterm: CS 6375 Spring 2015 Solutions Midterm: CS 6375 Spring 2015 Solutions The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for an

More information

Review of Lecture 1. Across records. Within records. Classification, Clustering, Outlier detection. Associations

Review of Lecture 1. Across records. Within records. Classification, Clustering, Outlier detection. Associations Review of Lecture 1 This course is about finding novel actionable patterns in data. We can divide data mining algorithms (and the patterns they find) into five groups Across records Classification, Clustering,

More information

Nonlinear Classification

Nonlinear Classification Nonlinear Classification INFO-4604, Applied Machine Learning University of Colorado Boulder October 5-10, 2017 Prof. Michael Paul Linear Classification Most classifiers we ve seen use linear functions

More information

Modeling Complex Temporal Composition of Actionlets for Activity Prediction

Modeling Complex Temporal Composition of Actionlets for Activity Prediction Modeling Complex Temporal Composition of Actionlets for Activity Prediction ECCV 2012 Activity Recognition Reading Group Framework of activity prediction What is an Actionlet To segment a long sequence

More information

Enabling Quality Control for Entity Resolution: A Human and Machine Cooperation Framework

Enabling Quality Control for Entity Resolution: A Human and Machine Cooperation Framework School of Computer Science, Northwestern Polytechnical University Enabling Quality Control for Entity Resolution: A Human and Machine Cooperation Framework Zhaoqiang Chen, Qun Chen, Fengfeng Fan, Yanyan

More information

CptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1

CptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1 CptS 570 Machine Learning School of EECS Washington State University CptS 570 - Machine Learning 1 IEEE Expert, October 1996 CptS 570 - Machine Learning 2 Given sample S from all possible examples D Learner

More information

Predicting flight on-time performance

Predicting flight on-time performance 1 Predicting flight on-time performance Arjun Mathur, Aaron Nagao, Kenny Ng I. INTRODUCTION Time is money, and delayed flights are a frequent cause of frustration for both travellers and airline companies.

More information

Anomaly Detection in Logged Sensor Data. Master s thesis in Complex Adaptive Systems JOHAN FLORBÄCK

Anomaly Detection in Logged Sensor Data. Master s thesis in Complex Adaptive Systems JOHAN FLORBÄCK Anomaly Detection in Logged Sensor Data Master s thesis in Complex Adaptive Systems JOHAN FLORBÄCK Department of Applied Mechanics CHALMERS UNIVERSITY OF TECHNOLOGY Göteborg, Sweden 2015 MASTER S THESIS

More information

.. Cal Poly CSC 466: Knowledge Discovery from Data Alexander Dekhtyar.. for each element of the dataset we are given its class label.

.. Cal Poly CSC 466: Knowledge Discovery from Data Alexander Dekhtyar.. for each element of the dataset we are given its class label. .. Cal Poly CSC 466: Knowledge Discovery from Data Alexander Dekhtyar.. Data Mining: Classification/Supervised Learning Definitions Data. Consider a set A = {A 1,...,A n } of attributes, and an additional

More information

Tennis player segmentation for semantic behavior analysis

Tennis player segmentation for semantic behavior analysis Proposta di Tennis player segmentation for semantic behavior analysis Architettura Software per Robot Mobili Vito Renò, Nicola Mosca, Massimiliano Nitti, Tiziana D Orazio, Donato Campagnoli, Andrea Prati,

More information

Spatial Decision Tree: A Novel Approach to Land-Cover Classification

Spatial Decision Tree: A Novel Approach to Land-Cover Classification Spatial Decision Tree: A Novel Approach to Land-Cover Classification Zhe Jiang 1, Shashi Shekhar 1, Xun Zhou 1, Joseph Knight 2, Jennifer Corcoran 2 1 Department of Computer Science & Engineering 2 Department

More information

Part I. Linear Discriminant Analysis. Discriminant analysis. Discriminant analysis

Part I. Linear Discriminant Analysis. Discriminant analysis. Discriminant analysis Week 5 Based in part on slides from textbook, slides of Susan Holmes Part I Linear Discriminant Analysis October 29, 2012 1 / 1 2 / 1 Nearest centroid rule Suppose we break down our data matrix as by the

More information

Global Scene Representations. Tilke Judd

Global Scene Representations. Tilke Judd Global Scene Representations Tilke Judd Papers Oliva and Torralba [2001] Fei Fei and Perona [2005] Labzebnik, Schmid and Ponce [2006] Commonalities Goal: Recognize natural scene categories Extract features

More information

Essence of Machine Learning (and Deep Learning) Hoa M. Le Data Science Lab, HUST hoamle.github.io

Essence of Machine Learning (and Deep Learning) Hoa M. Le Data Science Lab, HUST hoamle.github.io Essence of Machine Learning (and Deep Learning) Hoa M. Le Data Science Lab, HUST hoamle.github.io 1 Examples https://www.youtube.com/watch?v=bmka1zsg2 P4 http://www.r2d3.us/visual-intro-to-machinelearning-part-1/

More information

Performance Evaluation and Comparison

Performance Evaluation and Comparison Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation

More information

Ensemble Methods. NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan

Ensemble Methods. NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan Ensemble Methods NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan How do you make a decision? What do you want for lunch today?! What did you have last night?! What are your favorite

More information

Predictive Analytics on Accident Data Using Rule Based and Discriminative Classifiers

Predictive Analytics on Accident Data Using Rule Based and Discriminative Classifiers Advances in Computational Sciences and Technology ISSN 0973-6107 Volume 10, Number 3 (2017) pp. 461-469 Research India Publications http://www.ripublication.com Predictive Analytics on Accident Data Using

More information

Predictive analysis on Multivariate, Time Series datasets using Shapelets

Predictive analysis on Multivariate, Time Series datasets using Shapelets 1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,

More information

CS570 Data Mining. Anomaly Detection. Li Xiong. Slide credits: Tan, Steinbach, Kumar Jiawei Han and Micheline Kamber.

CS570 Data Mining. Anomaly Detection. Li Xiong. Slide credits: Tan, Steinbach, Kumar Jiawei Han and Micheline Kamber. CS570 Data Mining Anomaly Detection Li Xiong Slide credits: Tan, Steinbach, Kumar Jiawei Han and Micheline Kamber April 3, 2011 1 Anomaly Detection Anomaly is a pattern in the data that does not conform

More information

Bayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington

Bayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington Bayesian Classifiers and Probability Estimation Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington 1 Data Space Suppose that we have a classification problem The

More information

Machine Learning Linear Classification. Prof. Matteo Matteucci

Machine Learning Linear Classification. Prof. Matteo Matteucci Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)

More information

W vs. QCD Jet Tagging at the Large Hadron Collider

W vs. QCD Jet Tagging at the Large Hadron Collider W vs. QCD Jet Tagging at the Large Hadron Collider Bryan Anenberg: anenberg@stanford.edu; CS229 December 13, 2013 Problem Statement High energy collisions of protons at the Large Hadron Collider (LHC)

More information

MASTER. Anomaly detection on event logs an unsupervised algorithm on ixr-messages. Severins, J.D. Award date: Link to publication

MASTER. Anomaly detection on event logs an unsupervised algorithm on ixr-messages. Severins, J.D. Award date: Link to publication MASTER Anomaly detection on event logs an unsupervised algorithm on ixr-messages Severins, J.D. Award date: 2016 Link to publication Disclaimer This document contains a student thesis (bachelor's or master's),

More information

Some thoughts about anomaly detection in HEP

Some thoughts about anomaly detection in HEP Some thoughts about anomaly detection in HEP (with slides from M. Pierini and D. Rousseau) Vladimir V. Gligorov DS@HEP2016, Simons Foundation, New York What kind of anomaly? Will discuss two different

More information

Text Mining. Dr. Yanjun Li. Associate Professor. Department of Computer and Information Sciences Fordham University

Text Mining. Dr. Yanjun Li. Associate Professor. Department of Computer and Information Sciences Fordham University Text Mining Dr. Yanjun Li Associate Professor Department of Computer and Information Sciences Fordham University Outline Introduction: Data Mining Part One: Text Mining Part Two: Preprocessing Text Data

More information

Q1 (12 points): Chap 4 Exercise 3 (a) to (f) (2 points each)

Q1 (12 points): Chap 4 Exercise 3 (a) to (f) (2 points each) Q1 (1 points): Chap 4 Exercise 3 (a) to (f) ( points each) Given a table Table 1 Dataset for Exercise 3 Instance a 1 a a 3 Target Class 1 T T 1.0 + T T 6.0 + 3 T F 5.0-4 F F 4.0 + 5 F T 7.0-6 F T 3.0-7

More information

Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report

Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report Hujia Yu, Jiafu Wu [hujiay, jiafuwu]@stanford.edu 1. Introduction Housing prices are an important

More information

Smart Home Health Analytics Information Systems University of Maryland Baltimore County

Smart Home Health Analytics Information Systems University of Maryland Baltimore County Smart Home Health Analytics Information Systems University of Maryland Baltimore County 1 IEEE Expert, October 1996 2 Given sample S from all possible examples D Learner L learns hypothesis h based on

More information

Bayesian Networks Inference with Probabilistic Graphical Models

Bayesian Networks Inference with Probabilistic Graphical Models 4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning

More information

CS4445 Data Mining and Knowledge Discovery in Databases. B Term 2014 Solutions Exam 2 - December 15, 2014

CS4445 Data Mining and Knowledge Discovery in Databases. B Term 2014 Solutions Exam 2 - December 15, 2014 CS4445 Data Mining and Knowledge Discovery in Databases. B Term 2014 Solutions Exam 2 - December 15, 2014 Prof. Carolina Ruiz Department of Computer Science Worcester Polytechnic Institute NAME: Prof.

More information

P leiades: Subspace Clustering and Evaluation

P leiades: Subspace Clustering and Evaluation P leiades: Subspace Clustering and Evaluation Ira Assent, Emmanuel Müller, Ralph Krieger, Timm Jansen, and Thomas Seidl Data management and exploration group, RWTH Aachen University, Germany {assent,mueller,krieger,jansen,seidl}@cs.rwth-aachen.de

More information

Data Mining. Practical Machine Learning Tools and Techniques. Slides for Chapter 4 of Data Mining by I. H. Witten, E. Frank and M. A.

Data Mining. Practical Machine Learning Tools and Techniques. Slides for Chapter 4 of Data Mining by I. H. Witten, E. Frank and M. A. Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter of Data Mining by I. H. Witten, E. Frank and M. A. Hall Statistical modeling Opposite of R: use all the attributes Two assumptions:

More information

Short Introduction to CLIC and CTF3, Technologies for Future Linear Colliders

Short Introduction to CLIC and CTF3, Technologies for Future Linear Colliders Short Introduction to CLIC and CTF3, Technologies for Future Linear Colliders Explanation of the Basic Principles and Goals Visit to the CTF3 Installation Roger Ruber Collider History p p hadron collider

More information

Overview of IslandPick pipeline and the generation of GI datasets

Overview of IslandPick pipeline and the generation of GI datasets Overview of IslandPick pipeline and the generation of GI datasets Predicting GIs using comparative genomics By using whole genome alignments we can identify regions that are present in one genome but not

More information

Model Accuracy Measures

Model Accuracy Measures Model Accuracy Measures Master in Bioinformatics UPF 2017-2018 Eduardo Eyras Computational Genomics Pompeu Fabra University - ICREA Barcelona, Spain Variables What we can measure (attributes) Hypotheses

More information

FINAL: CS 6375 (Machine Learning) Fall 2014

FINAL: CS 6375 (Machine Learning) Fall 2014 FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for

More information

Data Mining Prof. Pabitra Mitra Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur

Data Mining Prof. Pabitra Mitra Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur Data Mining Prof. Pabitra Mitra Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur Lecture 21 K - Nearest Neighbor V In this lecture we discuss; how do we evaluate the

More information

Hunting for Anomalies in PMU Data

Hunting for Anomalies in PMU Data Hunting for Anomalies in PMU Data BRETT AMIDAN JAMES FOLLUM JEFFERY DAGLE Pacific Northwest National Laboratory NASPI Presentation (October 23, 2014) November 3, 2014 b.amidan@pnnl.gov 1 Big Picture Objective

More information

Stephen Scott.

Stephen Scott. 1 / 35 (Adapted from Ethem Alpaydin and Tom Mitchell) sscott@cse.unl.edu In Homework 1, you are (supposedly) 1 Choosing a data set 2 Extracting a test set of size > 30 3 Building a tree on the training

More information

Click Prediction and Preference Ranking of RSS Feeds

Click Prediction and Preference Ranking of RSS Feeds Click Prediction and Preference Ranking of RSS Feeds 1 Introduction December 11, 2009 Steven Wu RSS (Really Simple Syndication) is a family of data formats used to publish frequently updated works. RSS

More information

Tolerating Broken Robots

Tolerating Broken Robots Tolerating Broken Robots Lachlan Murray, Jon Timmis and Andy Tyrrell Intelligent Systems Group Department of Electronics University of York, UK March 24, 2011 People say my broken friend is useless, but

More information

Holdout and Cross-Validation Methods Overfitting Avoidance

Holdout and Cross-Validation Methods Overfitting Avoidance Holdout and Cross-Validation Methods Overfitting Avoidance Decision Trees Reduce error pruning Cost-complexity pruning Neural Networks Early stopping Adjusting Regularizers via Cross-Validation Nearest

More information

CSE 5243 INTRO. TO DATA MINING

CSE 5243 INTRO. TO DATA MINING CSE 5243 INTRO. TO DATA MINING Data & Data Preprocessing & Classification (Basic Concepts) Huan Sun, CSE@The Ohio State University Slides adapted from UIUC CS412, Fall 2017, by Prof. Jiawei Han Chapter

More information

Mejbah Alam. Justin Gottschlich. Tae Jun Lee. Stan Zdonik. Nesime Tatbul

Mejbah Alam. Justin Gottschlich. Tae Jun Lee. Stan Zdonik. Nesime Tatbul Nesime Tatbul Mejbah Alam Justin Gottschlich Tae Jun Lee Stan Zdonik 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montreal, Canada Motivation: Time Series Anomaly Detection

More information

Thesis. Wei Tang. 1 Abstract 3. 3 Experiment Background The Large Hadron Collider The ATLAS Detector... 4

Thesis. Wei Tang. 1 Abstract 3. 3 Experiment Background The Large Hadron Collider The ATLAS Detector... 4 Thesis Wei Tang Contents 1 Abstract 3 2 Introduction 3 3 Experiment Background 4 3.1 The Large Hadron Collider........................... 4 3.2 The ATLAS Detector.............................. 4 4 Search

More information

Brief Introduction of Machine Learning Techniques for Content Analysis

Brief Introduction of Machine Learning Techniques for Content Analysis 1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

New Prediction Methods for Tree Ensembles with Applications in Record Linkage

New Prediction Methods for Tree Ensembles with Applications in Record Linkage New Prediction Methods for Tree Ensembles with Applications in Record Linkage Samuel L. Ventura Rebecca Nugent Department of Statistics Carnegie Mellon University June 11, 2015 45th Symposium on the Interface

More information

Predicting New Search-Query Cluster Volume

Predicting New Search-Query Cluster Volume Predicting New Search-Query Cluster Volume Jacob Sisk, Cory Barr December 14, 2007 1 Problem Statement Search engines allow people to find information important to them, and search engine companies derive

More information

Least Squares Classification

Least Squares Classification Least Squares Classification Stephen Boyd EE103 Stanford University November 4, 2017 Outline Classification Least squares classification Multi-class classifiers Classification 2 Classification data fitting

More information

Lecture 7: DecisionTrees

Lecture 7: DecisionTrees Lecture 7: DecisionTrees What are decision trees? Brief interlude on information theory Decision tree construction Overfitting avoidance Regression trees COMP-652, Lecture 7 - September 28, 2009 1 Recall:

More information

International Journal of Scientific & Engineering Research, Volume 7, Issue 2, February ISSN

International Journal of Scientific & Engineering Research, Volume 7, Issue 2, February ISSN International Journal of Scientific & Engineering Research, Volume 7, Issue 2, February-2016 9 Automated Methodology for Context Based Semantic Anomaly Identification in Big Data Hema.R 1, Vidhya.V 2,

More information

Algorithmisches Lernen/Machine Learning

Algorithmisches Lernen/Machine Learning Algorithmisches Lernen/Machine Learning Part 1: Stefan Wermter Introduction Connectionist Learning (e.g. Neural Networks) Decision-Trees, Genetic Algorithms Part 2: Norman Hendrich Support-Vector Machines

More information

Statistical Filters for Crowd Image Analysis

Statistical Filters for Crowd Image Analysis Statistical Filters for Crowd Image Analysis Ákos Utasi, Ákos Kiss and Tamás Szirányi Distributed Events Analysis Research Group, Computer and Automation Research Institute H-1111 Budapest, Kende utca

More information

Machine Learning Lecture 5

Machine Learning Lecture 5 Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory

More information

Bayesian Decision Theory

Bayesian Decision Theory Introduction to Pattern Recognition [ Part 4 ] Mahdi Vasighi Remarks It is quite common to assume that the data in each class are adequately described by a Gaussian distribution. Bayesian classifier is

More information

Andras Hajdu Faculty of Informatics, University of Debrecen

Andras Hajdu Faculty of Informatics, University of Debrecen Ensemble-based based systems in medical image processing Andras Hajdu Faculty of Informatics, University of Debrecen SSIP 2011, Szeged, Hungary Ensemble based systems Ensemble learning is the process by

More information

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1 EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle

More information

A Framework for Adaptive Anomaly Detection Based on Support Vector Data Description

A Framework for Adaptive Anomaly Detection Based on Support Vector Data Description A Framework for Adaptive Anomaly Detection Based on Support Vector Data Description Min Yang, HuanGuo Zhang, JianMing Fu, and Fei Yan School of Computer, State Key Laboratory of Software Engineering, Wuhan

More information

Mitosis Detection in Breast Cancer Histology Images with Multi Column Deep Neural Networks

Mitosis Detection in Breast Cancer Histology Images with Multi Column Deep Neural Networks Mitosis Detection in Breast Cancer Histology Images with Multi Column Deep Neural Networks IDSIA, Lugano, Switzerland dan.ciresan@gmail.com Dan C. Cireşan and Alessandro Giusti DNN for Visual Pattern Recognition

More information

A Framework for Modeling Positive Class Expansion with Single Snapshot

A Framework for Modeling Positive Class Expansion with Single Snapshot A Framework for Modeling Positive Class Expansion with Single Snapshot Yang Yu and Zhi-Hua Zhou LAMDA Group National Key Laboratory for Novel Software Technology Nanjing University, China Motivating task

More information

ONR Mine Warfare Autonomy Virtual Program Review September 7, 2017

ONR Mine Warfare Autonomy Virtual Program Review September 7, 2017 ONR Mine Warfare Autonomy Virtual Program Review September 7, 2017 Information-driven Guidance and Control for Adaptive Target Detection and Classification Silvia Ferrari Pingping Zhu and Bo Fu Mechanical

More information

Application and Challenges of Artificial Intelligence in Exploration

Application and Challenges of Artificial Intelligence in Exploration Application and Challenges of Artificial Intelligence in Exploration XPLOR 2017 John McGaughey 2017 Mira Geoscience Ltd. Artificial Intelligence Artificial Intelligence is colossally hyped these days,

More information

Context-based Reasoning in Ambient Intelligence - CoReAmI -

Context-based Reasoning in Ambient Intelligence - CoReAmI - Context-based in Ambient Intelligence - CoReAmI - Hristijan Gjoreski Department of Intelligent Systems, Jožef Stefan Institute Supervisor: Prof. Dr. Matjaž Gams Co-supervisor: Dr. Mitja Luštrek Background

More information

the tree till a class assignment is reached

the tree till a class assignment is reached Decision Trees Decision Tree for Playing Tennis Prediction is done by sending the example down Prediction is done by sending the example down the tree till a class assignment is reached Definitions Internal

More information

Lars Schmidt-Thieme, Information Systems and Machine Learning Lab (ISMLL), University of Hildesheim, Germany

Lars Schmidt-Thieme, Information Systems and Machine Learning Lab (ISMLL), University of Hildesheim, Germany Syllabus Fri. 21.10. (1) 0. Introduction A. Supervised Learning: Linear Models & Fundamentals Fri. 27.10. (2) A.1 Linear Regression Fri. 3.11. (3) A.2 Linear Classification Fri. 10.11. (4) A.3 Regularization

More information

SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels

SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels Karl Stratos June 21, 2018 1 / 33 Tangent: Some Loose Ends in Logistic Regression Polynomial feature expansion in logistic

More information

Predicting Target Events in Industrial Domains

Predicting Target Events in Industrial Domains Predicting Target Events in Industrial Domains Julio Borges 1(B), Martin A. Neumann 1, Christian Bauer 2,YongDing 1, Till Riedel 1, and Michael Beigl 1 1 TECO, Karlsruhe Institute of Technology (KIT),

More information

Question of the Day. Machine Learning 2D1431. Decision Tree for PlayTennis. Outline. Lecture 4: Decision Tree Learning

Question of the Day. Machine Learning 2D1431. Decision Tree for PlayTennis. Outline. Lecture 4: Decision Tree Learning Question of the Day Machine Learning 2D1431 How can you make the following equation true by drawing only one straight line? 5 + 5 + 5 = 550 Lecture 4: Decision Tree Learning Outline Decision Tree for PlayTennis

More information

Temporal and spatial approaches for land cover classification.

Temporal and spatial approaches for land cover classification. Temporal and spatial approaches for land cover classification. Ryabukhin Sergey sergeyryabukhin@gmail.com Abstract. This paper describes solution for Time Series Land Cover Classification Challenge (TiSeLaC).

More information

Final Exam, Fall 2002

Final Exam, Fall 2002 15-781 Final Exam, Fall 22 1. Write your name and your andrew email address below. Name: Andrew ID: 2. There should be 17 pages in this exam (excluding this cover sheet). 3. If you need more room to work

More information

Introduction to Gaussian Process

Introduction to Gaussian Process Introduction to Gaussian Process CS 778 Chris Tensmeyer CS 478 INTRODUCTION 1 What Topic? Machine Learning Regression Bayesian ML Bayesian Regression Bayesian Non-parametric Gaussian Process (GP) GP Regression

More information

Classification and Prediction

Classification and Prediction Classification Classification and Prediction Classification: predict categorical class labels Build a model for a set of classes/concepts Classify loan applications (approve/decline) Prediction: model

More information

ECE 661: Homework 10 Fall 2014

ECE 661: Homework 10 Fall 2014 ECE 661: Homework 10 Fall 2014 This homework consists of the following two parts: (1) Face recognition with PCA and LDA for dimensionality reduction and the nearest-neighborhood rule for classification;

More information

Performance Evaluation

Performance Evaluation Performance Evaluation David S. Rosenberg Bloomberg ML EDU October 26, 2017 David S. Rosenberg (Bloomberg ML EDU) October 26, 2017 1 / 36 Baseline Models David S. Rosenberg (Bloomberg ML EDU) October 26,

More information

Robotics 2. AdaBoost for People and Place Detection. Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard

Robotics 2. AdaBoost for People and Place Detection. Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard Robotics 2 AdaBoost for People and Place Detection Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard v.1.1, Kai Arras, Jan 12, including material by Luciano Spinello and Oscar Martinez Mozos

More information

From statistics to data science. BAE 815 (Fall 2017) Dr. Zifei Liu

From statistics to data science. BAE 815 (Fall 2017) Dr. Zifei Liu From statistics to data science BAE 815 (Fall 2017) Dr. Zifei Liu Zifeiliu@ksu.edu Why? How? What? How much? How many? Individual facts (quantities, characters, or symbols) The Data-Information-Knowledge-Wisdom

More information

Induction of Decision Trees

Induction of Decision Trees Induction of Decision Trees Peter Waiganjo Wagacha This notes are for ICS320 Foundations of Learning and Adaptive Systems Institute of Computer Science University of Nairobi PO Box 30197, 00200 Nairobi.

More information

Machine Learning Analyses of Meteor Data

Machine Learning Analyses of Meteor Data WGN, The Journal of the IMO 45:5 (2017) 1 Machine Learning Analyses of Meteor Data Viswesh Krishna Research Student, Centre for Fundamental Research and Creative Education. Email: krishnaviswesh@cfrce.in

More information

Data Mining und Maschinelles Lernen

Data Mining und Maschinelles Lernen Data Mining und Maschinelles Lernen Ensemble Methods Bias-Variance Trade-off Basic Idea of Ensembles Bagging Basic Algorithm Bagging with Costs Randomization Random Forests Boosting Stacking Error-Correcting

More information

An Overview of Outlier Detection Techniques and Applications

An Overview of Outlier Detection Techniques and Applications Machine Learning Rhein-Neckar Meetup An Overview of Outlier Detection Techniques and Applications Ying Gu connygy@gmail.com 28.02.2016 Anomaly/Outlier Detection What are anomalies/outliers? The set of

More information

Introduction to Machine Learning Midterm Exam

Introduction to Machine Learning Midterm Exam 10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but

More information

Decision Trees. CS57300 Data Mining Fall Instructor: Bruno Ribeiro

Decision Trees. CS57300 Data Mining Fall Instructor: Bruno Ribeiro Decision Trees CS57300 Data Mining Fall 2016 Instructor: Bruno Ribeiro Goal } Classification without Models Well, partially without a model } Today: Decision Trees 2015 Bruno Ribeiro 2 3 Why Trees? } interpretable/intuitive,

More information

SYSTEMATIC CONSTRUCTION OF ANOMALY DETECTION BENCHMARKS FROM REAL DATA. Outlier Detection And Description Workshop 2013

SYSTEMATIC CONSTRUCTION OF ANOMALY DETECTION BENCHMARKS FROM REAL DATA. Outlier Detection And Description Workshop 2013 SYSTEMATIC CONSTRUCTION OF ANOMALY DETECTION BENCHMARKS FROM REAL DATA Outlier Detection And Description Workshop 2013 Authors Andrew Emmott emmott@eecs.oregonstate.edu Thomas Dietterich tgd@eecs.oregonstate.edu

More information