Anomaly Detection for the CERN Large Hadron Collider injection magnets
|
|
- Cathleen Walton
- 5 years ago
- Views:
Transcription
1 Anomaly Detection for the CERN Large Hadron Collider injection magnets Armin Halilovic KU Leuven - Department of Computer Science In cooperation with CERN
2 0 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 1 Anomaly Detection for the CERN Large Hadron Collider injection magnets
3 1 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 2 Anomaly Detection for the CERN Large Hadron Collider injection magnets
4 1 Context - Anomaly Detection Classification Normal vs. Abnormal/novel data One-class vs. Multiclass classification High amount of normal data Very low amount of anomalous data Unsupervised machine learning models Assign anomaly scores to data Outlier removal 3 Anomaly Detection for the CERN Large Hadron Collider injection magnets
5 1 Context - Problem Statement & Motivation The goal is to develop an anomaly detection application that can detect anomalies in the behaviour of the injection kicker magnets of the Large Hadron Collider. This is useful, because it can be used to: Detect anomalous behaviour and thus predict failures Improve CERN s response time Improve machine reliability 4 Anomaly Detection for the CERN Large Hadron Collider injection magnets
6 2 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 5 Anomaly Detection for the CERN Large Hadron Collider injection magnets
7 2 Data - Types I 6 types of data collections: 1 Continuous 2 Internal Post Operational Check (IPOC) 3 State 4 Controller 5 LHC 6 Electronic Logbook Continuous & discrete variables Fixed sampling rates & asynchronous sampling triggers 120 data collections Data from June 2015 to September Anomaly Detection for the CERN Large Hadron Collider injection magnets
8 2 Data - Types II Continuous Data: Temperature and pressures Fixed frequency sampling + save based on change in value Missing data: Forward Fill 7 Anomaly Detection for the CERN Large Hadron Collider injection magnets
9 2 Data - Types III Continuous Data: Temperature and pressures Fixed frequency sampling + save based on change in value Missing data: Forward Fill 8 Anomaly Detection for the CERN Large Hadron Collider injection magnets
10 2 Data - Types IV Internal Post Operational Check (IPOC) Data: Closely related to magnets: energy, strength, delay,... Only sampled when magnet generators pulse All IPOC measurements recorded simultaneously At most once every 10 seconds Many large gaps when experiments run Missing data: cannot fill Different timestamps for beams B1 and B2 Anomaly detection for the two MKI installations separately 9 Anomaly Detection for the CERN Large Hadron Collider injection magnets
11 2 Data - Types V IPOC, I STRENGTH, 2016: 10 Anomaly Detection for the CERN Large Hadron Collider injection magnets
12 2 Data - Types VI State Data: Not used No data for 2015 Controller Data: Not used Technical issues (duplicate timestamps) with received database 11 Anomaly Detection for the CERN Large Hadron Collider injection magnets
13 2 Data - Types VII LHC Data: Particle beam measurements: beam intensity & beam length Sampled and stored in similar way to Continuous measurements Missing data: Forward fill 12 Anomaly Detection for the CERN Large Hadron Collider injection magnets
14 2 Data - Types VIII Electronic Logbook Data: Manually created logbook entries (labels) Describe certain events Anomaly labels not precise, but range of 12 hours Label type Beam 1 Beam 2 anomaly fault info intervention research Total: Anomaly Detection for the CERN Large Hadron Collider injection magnets
15 2 Data - IPOC Segments I Magnets only in use for certain time periods IPOC data sampled only when magnets in use IPOC segment = period of magnet usage Introduced to deal with uncertainty of anomaly labels Important semantic meaning Data is split into segments based on segmentation distance 14 Anomaly Detection for the CERN Large Hadron Collider injection magnets
16 2 Data - IPOC Segments II Data is split into segments based on segmentation distance 15 Anomaly Detection for the CERN Large Hadron Collider injection magnets
17 3 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 16 Anomaly Detection for the CERN Large Hadron Collider injection magnets
18 3 Preprocessing - Data Filtering I Want to train models based on correct/relevant data Sudden extremely high temperatures, negative timing, etc. are impossible Measurement Minimum Maximum PRESSURE mbar mbar TEMP MAGNET (DOWN UP) 18 C 60 C TEMP TUBE (DOWN UP) 18 C 120 C I STRENGTH 1 ka N/A T DELAY 10 µs N/A 17 Anomaly Detection for the CERN Large Hadron Collider injection magnets
19 3 Preprocessing - Data Filtering II True pattern emerges 18 Anomaly Detection for the CERN Large Hadron Collider injection magnets
20 3 Preprocessing - Data Filtering III Impossible time delays removed 19 Anomaly Detection for the CERN Large Hadron Collider injection magnets
21 3 Preprocessing - Features All IPOC data + Continuous data at IPOC data timestamps (with forward fill) + LHC data at IPOC data timestamps (with forward fill) + Temporal features on Continuous and LHC data: To catch temporal relationship in data Sliding window features: mean & sum Important parameter: sliding window size Done separately for both B1 and B2 20 Anomaly Detection for the CERN Large Hadron Collider injection magnets
22 4 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 21 Anomaly Detection for the CERN Large Hadron Collider injection magnets
23 4 Anomaly Detection Train machine learning model using preprocessed data Use the model to generate anomaly scores Rescale scores to [0, 1] 22 Anomaly Detection for the CERN Large Hadron Collider injection magnets
24 4 Anomaly Detection - Isolation Forest Anomaly Scores 23 Anomaly Detection for the CERN Large Hadron Collider injection magnets
25 4 Anomaly Detection - Gaussian Mixture Model Scores I 24 Anomaly Detection for the CERN Large Hadron Collider injection magnets
26 4 Anomaly Detection - Gaussian Mixture Model Scores II 25 Anomaly Detection for the CERN Large Hadron Collider injection magnets
27 4 Anomaly Detection - Dummy Detectors Simple detection strategies as baseline to compare to Constant, uniformly random, stratified random 26 Anomaly Detection for the CERN Large Hadron Collider injection magnets
28 5 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 27 Anomaly Detection for the CERN Large Hadron Collider injection magnets
29 5 Postprocessing I Anomaly labels are unspecific, 12 hour range Will use segments instead of individual data tuples in evaluation Transform scored data into lists of IPOC segments Segment anomaly score based on anomaly scores of its data Anomalous behavior likely occurs in multiple successive timestamps These timestamps should get higher anomaly scores The segments that contain these timestamps should then have higher anomaly scores 28 Anomaly Detection for the CERN Large Hadron Collider injection magnets
30 5 Postprocessing II Methods for Segment Anomaly Score: Max Top K (10) Top Percentage (25%) Ground Truth Annotation: Need to compare segment anomaly scores to consistent basis of ground truth This allows for fair performance evaluation Mark segments as anomalous if they lie in the 12 hour range of an anomaly label 29 Anomaly Detection for the CERN Large Hadron Collider injection magnets
31 5 Postprocessing III We now have: A set of IPOC segments with anomaly scores Knowledge of which segments are actually anomalous 30 Anomaly Detection for the CERN Large Hadron Collider injection magnets
32 6 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 31 Anomaly Detection for the CERN Large Hadron Collider injection magnets
33 6 Evaluation Anomaly scores lie in [0, 1] Ground truth is 0 or 1 To evaluate performance, need to select a threshold anomaly score in order to count True Positives, False Positives, True Negatives, and False Negatives If score above threshold, then prediction is Positive, else Negative Prediction Ground Truth Positive Negative Positive TP FP Negative FN TN 32 Anomaly Detection for the CERN Large Hadron Collider injection magnets
34 6 Evaluation - Performance Metric Precision and Recall are useful context of imbalanced data P recision = T P T P +F P Recall = T P T P +F N But, want single number as performance metric for automated comparisons Calculate Precision and Recall for each possible anomaly score threshold and plot the resulting curve Performance metric = Area under Precision-Recall Curve (AUPR) 33 Anomaly Detection for the CERN Large Hadron Collider injection magnets
35 6 Evaluation - Grid Search Many parameters for developed anomaly detection pipeline Segmentation distance, scale data, anomaly score method, anomaly detector, anomaly detector hyperparameters, labels Grid search for parameter optimization Pipeline is executed automatically with predetermined combinations of parameters built by a grid of parameters Results are stored and sorted by AUPR so that the best performing parameters can be found easily 34 Anomaly Detection for the CERN Large Hadron Collider injection magnets
36 7 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 35 Anomaly Detection for the CERN Large Hadron Collider injection magnets
37 7 Results - Dummy Detectors Figure: PR curves of Dummy detectors with evaluation parameters segmentation distance = 30 min, anomaly score method = top k, labels = all. 36 Anomaly Detection for the CERN Large Hadron Collider injection magnets
38 7 Results - GMM I Best PR Curve Figure: Parameters: n components = 6, covariance type = full, scale data = F alse, segmentation distance = 60 min, anomaly score method = top k, labels = anomaly 37 Anomaly Detection for the CERN Large Hadron Collider injection magnets
39 7 Results - GMM II Predictions with 99-th percentile anomaly score threshold 38 Anomaly Detection for the CERN Large Hadron Collider injection magnets
40 7 Results - Isolation Forest I Best PR Curve Figure: Parameters: n estimators = 250, max samples = 5120, scale data = F alse, segmentation distance = 60 min, anomaly score method = max, labels = anomaly 39 Anomaly Detection for the CERN Large Hadron Collider injection magnets
41 7 Results - Isolation Forest II Predictions with 99-th percentile anomaly score threshold 40 Anomaly Detection for the CERN Large Hadron Collider injection magnets
42 7 Results - Isolation Forest III Only IPOC features, best PR Curve 41 Anomaly Detection for the CERN Large Hadron Collider injection magnets
43 7 Results - Isolation Forest IV Only IPOC features, predictions with 99-th percentile anomaly score threshold 42 Anomaly Detection for the CERN Large Hadron Collider injection magnets
44 7 Results - Source Code Written to be extensible Pipeline components in clear modules: preprocessing anomaly detection postprocessing evaluation pipeline Parameters can be varied easily 43 Anomaly Detection for the CERN Large Hadron Collider injection magnets
45 8 Outline 1 Context 2 Data 3 Preprocessing 4 Anomaly Detection 5 Postprocessing 6 Evaluation 7 Results 8 Conclusion 44 Anomaly Detection for the CERN Large Hadron Collider injection magnets
46 8 Conclusion Anomaly detection application has been developed Some anomalies are detected very well Many are still not detected at all Experiments have shown that performance can still be improved significantly More experiments should be done around feature selection 45 Anomaly Detection for the CERN Large Hadron Collider injection magnets
47 8 Future Work Feature selection Controller data Integration of more anomaly detectors (e.g. one class SVM or Local Outlier Factor) Better segmentation procedure without segmentation distance parameter More efficient and autonomous parameter optimization using e.g. Evolutionary algorithms or Bayesian Optimization 46 Anomaly Detection for the CERN Large Hadron Collider injection magnets
48 8 Bibliography CERN. Overview lhc. Accessed W Herr and T Pieloni. Beam-beam effects. (arxiv: ):1 29, Contribution to the CAS - CERN Accelerator School: Advanced Accelerator Physics Course, Trondheim, Norway, Aug Fei Tony Liu, Kai Ming Ting, and Zhi-Hua Zhou. Isolation forest. In Data Mining, ICDM 08. Eighth IEEE International Conference on, pages IEEE, Anomaly Detection for the CERN Large Hadron Collider injection magnets
49 Questions?
50 8 Extra - Comparison to Previous Work Enabled use of many machine learning models instead of just 1 Segmentation of input data instead of segmentation of output anomaly scores Consistent basis of ground truth more correct comparison of results Evaluation metrics in terms of TP, FP TN, FN instead of ambiguous terms PR curve using all anomaly score thresholds instead of calculating Precision and Recall for one threshold Anomaly Detection for the CERN Large Hadron Collider injection magnets
51 8 Extra - Isolation Forest Ensemble of simple decision trees which split randomly on features Trees are grown for random samples of dataset until each data tuple forms a leaf node Average path length will be shorter for anomalies Works well in high dimensional problems Density estimation, but without a density measure Source: [3] 49 Anomaly Detection for the CERN Large Hadron Collider injection magnets
Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts
Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 1 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification
More informationData Mining Based Anomaly Detection In PMU Measurements And Event Detection
Data Mining Based Anomaly Detection In PMU Measurements And Event Detection P. Banerjee, S. Pandey, M. Zhou, A. Srivastava, Y. Wu Smart Grid Demonstration and Research Investigation Lab (SGDRIL) Energy
More informationClass 4: Classification. Quaid Morris February 11 th, 2011 ML4Bio
Class 4: Classification Quaid Morris February 11 th, 211 ML4Bio Overview Basic concepts in classification: overfitting, cross-validation, evaluation. Linear Discriminant Analysis and Quadratic Discriminant
More informationMachine Learning Concepts in Chemoinformatics
Machine Learning Concepts in Chemoinformatics Martin Vogt B-IT Life Science Informatics Rheinische Friedrich-Wilhelms-Universität Bonn BigChem Winter School 2017 25. October Data Mining in Chemoinformatics
More informationCS145: INTRODUCTION TO DATA MINING
CS145: INTRODUCTION TO DATA MINING 4: Vector Data: Decision Tree Instructor: Yizhou Sun yzsun@cs.ucla.edu October 10, 2017 Methods to Learn Vector Data Set Data Sequence Data Text Data Classification Clustering
More informationEvaluation & Credibility Issues
Evaluation & Credibility Issues What measure should we use? accuracy might not be enough. How reliable are the predicted results? How much should we believe in what was learned? Error on the training data
More informationThe MD was done at 450GeV using beam 2 only. An MD focussing on injection of bunches with nominal emittance was done in parallel on beam 1.
CERN-ATS-Note-2011-065 MD 2011-08-08 Tobias.Baer@cern.ch MKI UFOs at Injection Tobias BAER, Mike BARNES, Wolfgang BARTMANN, Chiara BRACCO, Etienne CARLIER, Christophe CHANAVAT, Lene Norderhaug DROSDAL,
More informationEvaluation. Andrea Passerini Machine Learning. Evaluation
Andrea Passerini passerini@disi.unitn.it Machine Learning Basic concepts requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain
More informationAnomaly Detection. Jing Gao. SUNY Buffalo
Anomaly Detection Jing Gao SUNY Buffalo 1 Anomaly Detection Anomalies the set of objects are considerably dissimilar from the remainder of the data occur relatively infrequently when they do occur, their
More informationA New Unsupervised Event Detector for Non-Intrusive Load Monitoring
A New Unsupervised Event Detector for Non-Intrusive Load Monitoring GlobalSIP 2015, 14th Dec. Benjamin Wild, Karim Said Barsim, and Bin Yang Institute of Signal Processing and System Theory of,, Germany
More informationEvaluation requires to define performance measures to be optimized
Evaluation Basic concepts Evaluation requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain (generalization error) approximation
More informationSUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION
SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology
More informationUnsupervised Anomaly Detection for High Dimensional Data
Unsupervised Anomaly Detection for High Dimensional Data Department of Mathematics, Rowan University. July 19th, 2013 International Workshop in Sequential Methodologies (IWSM-2013) Outline of Talk Motivation
More informationLecture 3: Decision Trees
Lecture 3: Decision Trees Cognitive Systems - Machine Learning Part I: Basic Approaches of Concept Learning ID3, Information Gain, Overfitting, Pruning last change November 26, 2014 Ute Schmid (CogSys,
More information15-388/688 - Practical Data Science: Nonlinear modeling, cross-validation, regularization, and evaluation
15-388/688 - Practical Data Science: Nonlinear modeling, cross-validation, regularization, and evaluation J. Zico Kolter Carnegie Mellon University Fall 2016 1 Outline Example: return to peak demand prediction
More informationMidterm: CS 6375 Spring 2015 Solutions
Midterm: CS 6375 Spring 2015 Solutions The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for an
More informationReview of Lecture 1. Across records. Within records. Classification, Clustering, Outlier detection. Associations
Review of Lecture 1 This course is about finding novel actionable patterns in data. We can divide data mining algorithms (and the patterns they find) into five groups Across records Classification, Clustering,
More informationNonlinear Classification
Nonlinear Classification INFO-4604, Applied Machine Learning University of Colorado Boulder October 5-10, 2017 Prof. Michael Paul Linear Classification Most classifiers we ve seen use linear functions
More informationModeling Complex Temporal Composition of Actionlets for Activity Prediction
Modeling Complex Temporal Composition of Actionlets for Activity Prediction ECCV 2012 Activity Recognition Reading Group Framework of activity prediction What is an Actionlet To segment a long sequence
More informationEnabling Quality Control for Entity Resolution: A Human and Machine Cooperation Framework
School of Computer Science, Northwestern Polytechnical University Enabling Quality Control for Entity Resolution: A Human and Machine Cooperation Framework Zhaoqiang Chen, Qun Chen, Fengfeng Fan, Yanyan
More informationCptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1
CptS 570 Machine Learning School of EECS Washington State University CptS 570 - Machine Learning 1 IEEE Expert, October 1996 CptS 570 - Machine Learning 2 Given sample S from all possible examples D Learner
More informationPredicting flight on-time performance
1 Predicting flight on-time performance Arjun Mathur, Aaron Nagao, Kenny Ng I. INTRODUCTION Time is money, and delayed flights are a frequent cause of frustration for both travellers and airline companies.
More informationAnomaly Detection in Logged Sensor Data. Master s thesis in Complex Adaptive Systems JOHAN FLORBÄCK
Anomaly Detection in Logged Sensor Data Master s thesis in Complex Adaptive Systems JOHAN FLORBÄCK Department of Applied Mechanics CHALMERS UNIVERSITY OF TECHNOLOGY Göteborg, Sweden 2015 MASTER S THESIS
More information.. Cal Poly CSC 466: Knowledge Discovery from Data Alexander Dekhtyar.. for each element of the dataset we are given its class label.
.. Cal Poly CSC 466: Knowledge Discovery from Data Alexander Dekhtyar.. Data Mining: Classification/Supervised Learning Definitions Data. Consider a set A = {A 1,...,A n } of attributes, and an additional
More informationTennis player segmentation for semantic behavior analysis
Proposta di Tennis player segmentation for semantic behavior analysis Architettura Software per Robot Mobili Vito Renò, Nicola Mosca, Massimiliano Nitti, Tiziana D Orazio, Donato Campagnoli, Andrea Prati,
More informationSpatial Decision Tree: A Novel Approach to Land-Cover Classification
Spatial Decision Tree: A Novel Approach to Land-Cover Classification Zhe Jiang 1, Shashi Shekhar 1, Xun Zhou 1, Joseph Knight 2, Jennifer Corcoran 2 1 Department of Computer Science & Engineering 2 Department
More informationPart I. Linear Discriminant Analysis. Discriminant analysis. Discriminant analysis
Week 5 Based in part on slides from textbook, slides of Susan Holmes Part I Linear Discriminant Analysis October 29, 2012 1 / 1 2 / 1 Nearest centroid rule Suppose we break down our data matrix as by the
More informationGlobal Scene Representations. Tilke Judd
Global Scene Representations Tilke Judd Papers Oliva and Torralba [2001] Fei Fei and Perona [2005] Labzebnik, Schmid and Ponce [2006] Commonalities Goal: Recognize natural scene categories Extract features
More informationEssence of Machine Learning (and Deep Learning) Hoa M. Le Data Science Lab, HUST hoamle.github.io
Essence of Machine Learning (and Deep Learning) Hoa M. Le Data Science Lab, HUST hoamle.github.io 1 Examples https://www.youtube.com/watch?v=bmka1zsg2 P4 http://www.r2d3.us/visual-intro-to-machinelearning-part-1/
More informationPerformance Evaluation and Comparison
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation
More informationEnsemble Methods. NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan
Ensemble Methods NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan How do you make a decision? What do you want for lunch today?! What did you have last night?! What are your favorite
More informationPredictive Analytics on Accident Data Using Rule Based and Discriminative Classifiers
Advances in Computational Sciences and Technology ISSN 0973-6107 Volume 10, Number 3 (2017) pp. 461-469 Research India Publications http://www.ripublication.com Predictive Analytics on Accident Data Using
More informationPredictive analysis on Multivariate, Time Series datasets using Shapelets
1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,
More informationCS570 Data Mining. Anomaly Detection. Li Xiong. Slide credits: Tan, Steinbach, Kumar Jiawei Han and Micheline Kamber.
CS570 Data Mining Anomaly Detection Li Xiong Slide credits: Tan, Steinbach, Kumar Jiawei Han and Micheline Kamber April 3, 2011 1 Anomaly Detection Anomaly is a pattern in the data that does not conform
More informationBayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington
Bayesian Classifiers and Probability Estimation Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington 1 Data Space Suppose that we have a classification problem The
More informationMachine Learning Linear Classification. Prof. Matteo Matteucci
Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)
More informationW vs. QCD Jet Tagging at the Large Hadron Collider
W vs. QCD Jet Tagging at the Large Hadron Collider Bryan Anenberg: anenberg@stanford.edu; CS229 December 13, 2013 Problem Statement High energy collisions of protons at the Large Hadron Collider (LHC)
More informationMASTER. Anomaly detection on event logs an unsupervised algorithm on ixr-messages. Severins, J.D. Award date: Link to publication
MASTER Anomaly detection on event logs an unsupervised algorithm on ixr-messages Severins, J.D. Award date: 2016 Link to publication Disclaimer This document contains a student thesis (bachelor's or master's),
More informationSome thoughts about anomaly detection in HEP
Some thoughts about anomaly detection in HEP (with slides from M. Pierini and D. Rousseau) Vladimir V. Gligorov DS@HEP2016, Simons Foundation, New York What kind of anomaly? Will discuss two different
More informationText Mining. Dr. Yanjun Li. Associate Professor. Department of Computer and Information Sciences Fordham University
Text Mining Dr. Yanjun Li Associate Professor Department of Computer and Information Sciences Fordham University Outline Introduction: Data Mining Part One: Text Mining Part Two: Preprocessing Text Data
More informationQ1 (12 points): Chap 4 Exercise 3 (a) to (f) (2 points each)
Q1 (1 points): Chap 4 Exercise 3 (a) to (f) ( points each) Given a table Table 1 Dataset for Exercise 3 Instance a 1 a a 3 Target Class 1 T T 1.0 + T T 6.0 + 3 T F 5.0-4 F F 4.0 + 5 F T 7.0-6 F T 3.0-7
More informationReal Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report
Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report Hujia Yu, Jiafu Wu [hujiay, jiafuwu]@stanford.edu 1. Introduction Housing prices are an important
More informationSmart Home Health Analytics Information Systems University of Maryland Baltimore County
Smart Home Health Analytics Information Systems University of Maryland Baltimore County 1 IEEE Expert, October 1996 2 Given sample S from all possible examples D Learner L learns hypothesis h based on
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationCS4445 Data Mining and Knowledge Discovery in Databases. B Term 2014 Solutions Exam 2 - December 15, 2014
CS4445 Data Mining and Knowledge Discovery in Databases. B Term 2014 Solutions Exam 2 - December 15, 2014 Prof. Carolina Ruiz Department of Computer Science Worcester Polytechnic Institute NAME: Prof.
More informationP leiades: Subspace Clustering and Evaluation
P leiades: Subspace Clustering and Evaluation Ira Assent, Emmanuel Müller, Ralph Krieger, Timm Jansen, and Thomas Seidl Data management and exploration group, RWTH Aachen University, Germany {assent,mueller,krieger,jansen,seidl}@cs.rwth-aachen.de
More informationData Mining. Practical Machine Learning Tools and Techniques. Slides for Chapter 4 of Data Mining by I. H. Witten, E. Frank and M. A.
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter of Data Mining by I. H. Witten, E. Frank and M. A. Hall Statistical modeling Opposite of R: use all the attributes Two assumptions:
More informationShort Introduction to CLIC and CTF3, Technologies for Future Linear Colliders
Short Introduction to CLIC and CTF3, Technologies for Future Linear Colliders Explanation of the Basic Principles and Goals Visit to the CTF3 Installation Roger Ruber Collider History p p hadron collider
More informationOverview of IslandPick pipeline and the generation of GI datasets
Overview of IslandPick pipeline and the generation of GI datasets Predicting GIs using comparative genomics By using whole genome alignments we can identify regions that are present in one genome but not
More informationModel Accuracy Measures
Model Accuracy Measures Master in Bioinformatics UPF 2017-2018 Eduardo Eyras Computational Genomics Pompeu Fabra University - ICREA Barcelona, Spain Variables What we can measure (attributes) Hypotheses
More informationFINAL: CS 6375 (Machine Learning) Fall 2014
FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for
More informationData Mining Prof. Pabitra Mitra Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur
Data Mining Prof. Pabitra Mitra Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur Lecture 21 K - Nearest Neighbor V In this lecture we discuss; how do we evaluate the
More informationHunting for Anomalies in PMU Data
Hunting for Anomalies in PMU Data BRETT AMIDAN JAMES FOLLUM JEFFERY DAGLE Pacific Northwest National Laboratory NASPI Presentation (October 23, 2014) November 3, 2014 b.amidan@pnnl.gov 1 Big Picture Objective
More informationStephen Scott.
1 / 35 (Adapted from Ethem Alpaydin and Tom Mitchell) sscott@cse.unl.edu In Homework 1, you are (supposedly) 1 Choosing a data set 2 Extracting a test set of size > 30 3 Building a tree on the training
More informationClick Prediction and Preference Ranking of RSS Feeds
Click Prediction and Preference Ranking of RSS Feeds 1 Introduction December 11, 2009 Steven Wu RSS (Really Simple Syndication) is a family of data formats used to publish frequently updated works. RSS
More informationTolerating Broken Robots
Tolerating Broken Robots Lachlan Murray, Jon Timmis and Andy Tyrrell Intelligent Systems Group Department of Electronics University of York, UK March 24, 2011 People say my broken friend is useless, but
More informationHoldout and Cross-Validation Methods Overfitting Avoidance
Holdout and Cross-Validation Methods Overfitting Avoidance Decision Trees Reduce error pruning Cost-complexity pruning Neural Networks Early stopping Adjusting Regularizers via Cross-Validation Nearest
More informationCSE 5243 INTRO. TO DATA MINING
CSE 5243 INTRO. TO DATA MINING Data & Data Preprocessing & Classification (Basic Concepts) Huan Sun, CSE@The Ohio State University Slides adapted from UIUC CS412, Fall 2017, by Prof. Jiawei Han Chapter
More informationMejbah Alam. Justin Gottschlich. Tae Jun Lee. Stan Zdonik. Nesime Tatbul
Nesime Tatbul Mejbah Alam Justin Gottschlich Tae Jun Lee Stan Zdonik 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montreal, Canada Motivation: Time Series Anomaly Detection
More informationThesis. Wei Tang. 1 Abstract 3. 3 Experiment Background The Large Hadron Collider The ATLAS Detector... 4
Thesis Wei Tang Contents 1 Abstract 3 2 Introduction 3 3 Experiment Background 4 3.1 The Large Hadron Collider........................... 4 3.2 The ATLAS Detector.............................. 4 4 Search
More informationBrief Introduction of Machine Learning Techniques for Content Analysis
1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview
More informationMachine learning for pervasive systems Classification in high-dimensional spaces
Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version
More informationNew Prediction Methods for Tree Ensembles with Applications in Record Linkage
New Prediction Methods for Tree Ensembles with Applications in Record Linkage Samuel L. Ventura Rebecca Nugent Department of Statistics Carnegie Mellon University June 11, 2015 45th Symposium on the Interface
More informationPredicting New Search-Query Cluster Volume
Predicting New Search-Query Cluster Volume Jacob Sisk, Cory Barr December 14, 2007 1 Problem Statement Search engines allow people to find information important to them, and search engine companies derive
More informationLeast Squares Classification
Least Squares Classification Stephen Boyd EE103 Stanford University November 4, 2017 Outline Classification Least squares classification Multi-class classifiers Classification 2 Classification data fitting
More informationLecture 7: DecisionTrees
Lecture 7: DecisionTrees What are decision trees? Brief interlude on information theory Decision tree construction Overfitting avoidance Regression trees COMP-652, Lecture 7 - September 28, 2009 1 Recall:
More informationInternational Journal of Scientific & Engineering Research, Volume 7, Issue 2, February ISSN
International Journal of Scientific & Engineering Research, Volume 7, Issue 2, February-2016 9 Automated Methodology for Context Based Semantic Anomaly Identification in Big Data Hema.R 1, Vidhya.V 2,
More informationAlgorithmisches Lernen/Machine Learning
Algorithmisches Lernen/Machine Learning Part 1: Stefan Wermter Introduction Connectionist Learning (e.g. Neural Networks) Decision-Trees, Genetic Algorithms Part 2: Norman Hendrich Support-Vector Machines
More informationStatistical Filters for Crowd Image Analysis
Statistical Filters for Crowd Image Analysis Ákos Utasi, Ákos Kiss and Tamás Szirányi Distributed Events Analysis Research Group, Computer and Automation Research Institute H-1111 Budapest, Kende utca
More informationMachine Learning Lecture 5
Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory
More informationBayesian Decision Theory
Introduction to Pattern Recognition [ Part 4 ] Mahdi Vasighi Remarks It is quite common to assume that the data in each class are adequately described by a Gaussian distribution. Bayesian classifier is
More informationAndras Hajdu Faculty of Informatics, University of Debrecen
Ensemble-based based systems in medical image processing Andras Hajdu Faculty of Informatics, University of Debrecen SSIP 2011, Szeged, Hungary Ensemble based systems Ensemble learning is the process by
More informationEEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1
EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle
More informationA Framework for Adaptive Anomaly Detection Based on Support Vector Data Description
A Framework for Adaptive Anomaly Detection Based on Support Vector Data Description Min Yang, HuanGuo Zhang, JianMing Fu, and Fei Yan School of Computer, State Key Laboratory of Software Engineering, Wuhan
More informationMitosis Detection in Breast Cancer Histology Images with Multi Column Deep Neural Networks
Mitosis Detection in Breast Cancer Histology Images with Multi Column Deep Neural Networks IDSIA, Lugano, Switzerland dan.ciresan@gmail.com Dan C. Cireşan and Alessandro Giusti DNN for Visual Pattern Recognition
More informationA Framework for Modeling Positive Class Expansion with Single Snapshot
A Framework for Modeling Positive Class Expansion with Single Snapshot Yang Yu and Zhi-Hua Zhou LAMDA Group National Key Laboratory for Novel Software Technology Nanjing University, China Motivating task
More informationONR Mine Warfare Autonomy Virtual Program Review September 7, 2017
ONR Mine Warfare Autonomy Virtual Program Review September 7, 2017 Information-driven Guidance and Control for Adaptive Target Detection and Classification Silvia Ferrari Pingping Zhu and Bo Fu Mechanical
More informationApplication and Challenges of Artificial Intelligence in Exploration
Application and Challenges of Artificial Intelligence in Exploration XPLOR 2017 John McGaughey 2017 Mira Geoscience Ltd. Artificial Intelligence Artificial Intelligence is colossally hyped these days,
More informationContext-based Reasoning in Ambient Intelligence - CoReAmI -
Context-based in Ambient Intelligence - CoReAmI - Hristijan Gjoreski Department of Intelligent Systems, Jožef Stefan Institute Supervisor: Prof. Dr. Matjaž Gams Co-supervisor: Dr. Mitja Luštrek Background
More informationthe tree till a class assignment is reached
Decision Trees Decision Tree for Playing Tennis Prediction is done by sending the example down Prediction is done by sending the example down the tree till a class assignment is reached Definitions Internal
More informationLars Schmidt-Thieme, Information Systems and Machine Learning Lab (ISMLL), University of Hildesheim, Germany
Syllabus Fri. 21.10. (1) 0. Introduction A. Supervised Learning: Linear Models & Fundamentals Fri. 27.10. (2) A.1 Linear Regression Fri. 3.11. (3) A.2 Linear Classification Fri. 10.11. (4) A.3 Regularization
More informationSVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels
SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels Karl Stratos June 21, 2018 1 / 33 Tangent: Some Loose Ends in Logistic Regression Polynomial feature expansion in logistic
More informationPredicting Target Events in Industrial Domains
Predicting Target Events in Industrial Domains Julio Borges 1(B), Martin A. Neumann 1, Christian Bauer 2,YongDing 1, Till Riedel 1, and Michael Beigl 1 1 TECO, Karlsruhe Institute of Technology (KIT),
More informationQuestion of the Day. Machine Learning 2D1431. Decision Tree for PlayTennis. Outline. Lecture 4: Decision Tree Learning
Question of the Day Machine Learning 2D1431 How can you make the following equation true by drawing only one straight line? 5 + 5 + 5 = 550 Lecture 4: Decision Tree Learning Outline Decision Tree for PlayTennis
More informationTemporal and spatial approaches for land cover classification.
Temporal and spatial approaches for land cover classification. Ryabukhin Sergey sergeyryabukhin@gmail.com Abstract. This paper describes solution for Time Series Land Cover Classification Challenge (TiSeLaC).
More informationFinal Exam, Fall 2002
15-781 Final Exam, Fall 22 1. Write your name and your andrew email address below. Name: Andrew ID: 2. There should be 17 pages in this exam (excluding this cover sheet). 3. If you need more room to work
More informationIntroduction to Gaussian Process
Introduction to Gaussian Process CS 778 Chris Tensmeyer CS 478 INTRODUCTION 1 What Topic? Machine Learning Regression Bayesian ML Bayesian Regression Bayesian Non-parametric Gaussian Process (GP) GP Regression
More informationClassification and Prediction
Classification Classification and Prediction Classification: predict categorical class labels Build a model for a set of classes/concepts Classify loan applications (approve/decline) Prediction: model
More informationECE 661: Homework 10 Fall 2014
ECE 661: Homework 10 Fall 2014 This homework consists of the following two parts: (1) Face recognition with PCA and LDA for dimensionality reduction and the nearest-neighborhood rule for classification;
More informationPerformance Evaluation
Performance Evaluation David S. Rosenberg Bloomberg ML EDU October 26, 2017 David S. Rosenberg (Bloomberg ML EDU) October 26, 2017 1 / 36 Baseline Models David S. Rosenberg (Bloomberg ML EDU) October 26,
More informationRobotics 2. AdaBoost for People and Place Detection. Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard
Robotics 2 AdaBoost for People and Place Detection Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard v.1.1, Kai Arras, Jan 12, including material by Luciano Spinello and Oscar Martinez Mozos
More informationFrom statistics to data science. BAE 815 (Fall 2017) Dr. Zifei Liu
From statistics to data science BAE 815 (Fall 2017) Dr. Zifei Liu Zifeiliu@ksu.edu Why? How? What? How much? How many? Individual facts (quantities, characters, or symbols) The Data-Information-Knowledge-Wisdom
More informationInduction of Decision Trees
Induction of Decision Trees Peter Waiganjo Wagacha This notes are for ICS320 Foundations of Learning and Adaptive Systems Institute of Computer Science University of Nairobi PO Box 30197, 00200 Nairobi.
More informationMachine Learning Analyses of Meteor Data
WGN, The Journal of the IMO 45:5 (2017) 1 Machine Learning Analyses of Meteor Data Viswesh Krishna Research Student, Centre for Fundamental Research and Creative Education. Email: krishnaviswesh@cfrce.in
More informationData Mining und Maschinelles Lernen
Data Mining und Maschinelles Lernen Ensemble Methods Bias-Variance Trade-off Basic Idea of Ensembles Bagging Basic Algorithm Bagging with Costs Randomization Random Forests Boosting Stacking Error-Correcting
More informationAn Overview of Outlier Detection Techniques and Applications
Machine Learning Rhein-Neckar Meetup An Overview of Outlier Detection Techniques and Applications Ying Gu connygy@gmail.com 28.02.2016 Anomaly/Outlier Detection What are anomalies/outliers? The set of
More informationIntroduction to Machine Learning Midterm Exam
10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but
More informationDecision Trees. CS57300 Data Mining Fall Instructor: Bruno Ribeiro
Decision Trees CS57300 Data Mining Fall 2016 Instructor: Bruno Ribeiro Goal } Classification without Models Well, partially without a model } Today: Decision Trees 2015 Bruno Ribeiro 2 3 Why Trees? } interpretable/intuitive,
More informationSYSTEMATIC CONSTRUCTION OF ANOMALY DETECTION BENCHMARKS FROM REAL DATA. Outlier Detection And Description Workshop 2013
SYSTEMATIC CONSTRUCTION OF ANOMALY DETECTION BENCHMARKS FROM REAL DATA Outlier Detection And Description Workshop 2013 Authors Andrew Emmott emmott@eecs.oregonstate.edu Thomas Dietterich tgd@eecs.oregonstate.edu
More information