Machine Learning Analyses of Meteor Data

Similar documents
Temporal and spatial approaches for land cover classification.

Multi-Plant Photovoltaic Energy Forecasting Challenge: Second place solution

Searching for Long-Period Comets with Deep Learning Tools

Leveraging Machine Learning for High-Resolution Restoration of Satellite Imagery

PCA and Autoencoders

Planet Labels - How do we use our planet?

De-biasing of the velocity determination for double station meteor observations from CILBO

CS 229 Final Report: Data-Driven Prediction of Band Gap of Materials

arxiv: v2 [cs.lg] 5 May 2015

Training neural networks for financial forecasting: Backpropagation vs Particle Swarm Optimization

What s Cooking? Predicting Cuisines from Recipe Ingredients

Visual observations of Geminid meteor shower 2004

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

On the Interpretability of Conditional Probability Estimates in the Agnostic Setting

Term Paper: How and Why to Use Stochastic Gradient Descent?

arxiv: v3 [cs.lg] 23 Nov 2016

Yo Home to Bel-Air: Predicting Crime on The Streets of Philadelphia

Electronic Supporting Information Topological design of porous organic molecules

Characterizing Interference in Radio Astronomy Observations through Active and Unsupervised Learning

CPSC 340: Machine Learning and Data Mining. Linear Least Squares Fall 2016

Predictive Analytics on Accident Data Using Rule Based and Discriminative Classifiers

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

MST radar observations of the Leonid meteor storm during

Matrix Factorization for Speech Enhancement

Anomaly Detection for the CERN Large Hadron Collider injection magnets

Probabilistic Traffic Models for Occupancy Counting

Abstention Protocol for Accuracy and Speed

Text Mining. Dr. Yanjun Li. Associate Professor. Department of Computer and Information Sciences Fordham University

Anomaly Detection in Logged Sensor Data. Master s thesis in Complex Adaptive Systems JOHAN FLORBÄCK

INTRODUCTION TO DATA SCIENCE

Neural Networks Architecture Evaluation in a Quantum Computer

MATDAT18: Materials and Data Science Hackathon. Name Department Institution Jessica Kong Chemistry University of Washington

Unsupervised Anomaly Detection for High Dimensional Data

Analyzing the Earth Using Remote Sensing

Making Our Cities Safer: A Study In Neighbhorhood Crime Patterns

SEA surface temperature, SST for short, is an important

Predictive analysis on Multivariate, Time Series datasets using Shapelets

CS570 Introduction to Data Mining

Supplemental Information: A universal reaction coordinate for azobenzene photoisomerization

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Introduction. Chapter 1

IMPROVING INFRASTRUCTURE FOR TRANSPORTATION SYSTEMS USING CLUSTERING

Lectures in AstroStatistics: Topics in Machine Learning for Astronomers

W vs. QCD Jet Tagging at the Large Hadron Collider

Chart types and when to use them

CS6375: Machine Learning Gautam Kunapuli. Decision Trees

Weather Prediction Using Historical Data

arxiv: v1 [stat.ml] 12 May 2016

Classification with Perceptrons. Reading:

arxiv: v1 [stat.ml] 6 Nov 2018

A Community Gridded Atmospheric Forecast System for Calibrated Solar Irradiance

Statistical Machine Learning Theory. From Multi-class Classification to Structured Output Prediction. Hisashi Kashima.

Course in Data Science

Final Exam, Fall 2002

Data Mining Classification: Basic Concepts and Techniques. Lecture Notes for Chapter 3. Introduction to Data Mining, 2nd Edition

DATA SCIENCE SIMPLIFIED USING ARCGIS API FOR PYTHON

The Structure and Timescales of Heat Perception in Larval Zebrafish

Logistic Regression. COMP 527 Danushka Bollegala

Use of reanalysis data in atmospheric electricity studies

CLOUD NOWCASTING: MOTION ANALYSIS OF ALL-SKY IMAGES USING VELOCITY FIELDS

Learning from Examples

126 WGN, the Journal of the IMO 39:5 (2011)

Recognizing Actions during Tactile Manipulations through Force Sensing

Observing Dark Worlds (Final Report)

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report

Development of Orbit Analysis System for Spaceguard

Style-aware Mid-level Representation for Discovering Visual Connections in Space and Time

Fundamentals of meteor science

Department of Computer Science, University of Waikato, Hamilton, New Zealand. 2

Confirmation of the April Rho Cygnids (ARC, IAU#348)

CLUe Training An Introduction to Machine Learning in R with an example from handwritten digit recognition

Introduction to Machine Learning Midterm Exam

Predicting flight on-time performance

Retrieval of Cloud Top Pressure

The Perceptron Algorithm

Intelligence and statistics for rapid and robust earthquake detection, association and location

Jeffrey D. Ullman Stanford University

Quark-gluon tagging in the forward region of ATLAS at the LHC.

Machine Learning Lecture 6 Note

CSC Neural Networks. Perceptron Learning Rule

From Binary to Multiclass Classification. CS 6961: Structured Prediction Spring 2018

ECE521 W17 Tutorial 1. Renjie Liao & Min Bai

Doing Right By Massive Data: How To Bring Probability Modeling To The Analysis Of Huge Datasets Without Taking Over The Datacenter

Relate Attributes and Counts

Data-driven Avalanche Forecasting

Interpreting Deep Classifiers

Tufts COMP 135: Introduction to Machine Learning

Evolutionary Functional Link Interval Type-2 Fuzzy Neural System for Exchange Rate Prediction

Outline Challenges of Massive Data Combining approaches Application: Event Detection for Astronomical Data Conclusion. Abstract

Midterm. You may use a calculator, but not any device that can access the Internet or store large amounts of data.

@SoyGema GEMA PARREÑO PIQUERAS

Decision Trees. CS57300 Data Mining Fall Instructor: Bruno Ribeiro

2MHR. Protein structure classification is important because it organizes the protein structure universe that is independent of sequence similarity.

Combing Open-Source Programming Languages with GIS for Spatial Data Science. Maja Kalinic Master s Thesis

Unit 3 and 4 Further Mathematics: Exam 2

Hot spot Analysis with Clustering on NIJ Data Set The Data Set:

Statistical Machine Learning Theory. From Multi-class Classification to Structured Output Prediction. Hisashi Kashima.

A new method of discovering new meteor showers from the IMO single-station video meteor database

Decision trees COMS 4771

Transcription:

WGN, The Journal of the IMO 45:5 (2017) 1 Machine Learning Analyses of Meteor Data Viswesh Krishna Research Student, Centre for Fundamental Research and Creative Education. Email: krishnaviswesh@cfrce.in

2 WGN, The Journal of the IMO 45:5 (2017) The objective of this paper is to analyse meteor data through statistical analysis and machine learning. Two examples are used to highlight the effectiveness of these tools in the analysis of meteor data : Outlier Detection and Feature Prediction. Outlier Detection consists of identifying meteor outbursts from the data using statistical methods. Feature Prediction involves predicting the Level, Shower Type and Next Occurrence of a predicted meteor shower with machine learning models. Code is available at https://github.com/visweshk/outburstdetection-featureprediction 1 Introduction Meteor shower forecasting is today done through huge servers and supercomputers by plotting the exact trajectories of various objects in space and calculating the time to strike the atmosphere. However, this prediction is fraught with uncertainty due to varying levels of accuracy in the measurements of the position and velocity of the astronomical objects under consideration (Vaubaillon, 2017). Further, it is impossible to predict other features of a meteor shower before hand as these do not correspond to any concrete mathematical formulae. It is known that large datasets of meteor shower activity are uploaded by the International Meteor Organisation, NASA and other organisations. With the fast growth in Machine Learning and Data Analytics over the past few years, it is possible that new models could highlight previously unnoticed trends in meteor data. Indeed, as most of the meteor shower prediction is based on the past behaviour of the shower and parent bodies, it is clear that the application of these newly developed tools could provide insight into the relationship between meteor shower features. Machine Learning works with data and processes it to discover patterns that can be later used to analyse new data. Machine learning usually relies on a specific representation of the data; a set of features that are understandable for a computer. For the current project, Supervised learning was applied. Supervised machine learning algorithms can apply what has been learned in the past to new data using labeled examples to predict future events. Starting from the analysis of a known training dataset, the learning algorithm produces an inferred function to make predictions about the output values. The system is able to provide targets for any new input after sufficient training. The learning algorithm can also compare its output with the correct, intended output and find errors in order to modify the model accordingly. With the availability of 35 years of IMO meteor data, it is possible to build a supervised learning model for predicting future events. In this paper, we will explore some of the models which can be applied for supervised learning and their prediction ability. 2 General Preprocessing and Methods The datasets were downloaded from the International Meteor Organisation (IMO) Visual Meteor DataBase (IMO, 1982), consisting of 35 years of meteor data. The columns (features) in the database were User ID, Start Date, End Date, Right Ascension, Declination, Teff, F, Lm, Shower Type, Method, Number, Latitude, Longitude and Elevation (of the observatory). The processing was first done on R. The MetFns package was developed by the IMO on R (Veljkovic, 2017) and allowed for the generation of new features such as Solar Longitude, and Zenithal Hourly Rates to enable better prediction. 2.1 Processing on R The preprocessing of the data was first done on R as it enabled easily manipulation of the datasets and the MetFns package developed by the IMO could also be accessed here. The data was extracted from the CSV files, features were scaled and unnecessary indexing features were removed. Removal of duplicate entries was accomplished by using the differences in the Start Date, End Date, Right Ascension, Declination and Shower Type values. 2.2 Processing on Python On Python, the data was extracted from the CSV file through the Pandas (McKinney, 2010) module. For scaling the data, the Standard Scaler and MinMaxScaler in the Preprocessing module of Scikit-Learn (Pedregosa et al., 2011) were utilised to normalize the features to advised values ([-1,+1]) for the predictive models. Label Encoders from the Scikit-Learn module were used to index the various strings in the Shower data and convert them to integer values as string values could not be processed by certain machine learning models. Further, the latitude, longitude and elevation of an observatory were collapsed into a single index through another label encoder for ease of classification. The Keras (Chollet et al., 2015) and Scikit-Learn libraries on Python were used for building and evaluating the machine learning models. The standard training vs testing split for the supervised learning was set at 80-20 : 80% of the data was used to train the model and 20% was used to test the model. 3 Outlier Detection 3.1 Introduction Outlier Detection involved the detection of meteor outbursts from the dataset through statistical analysis. Me-

WGN, The Journal of the IMO 45:5 (2017) 3 teor outbursts could be thought of as irregular meteor showers with a high ZHR value and could be detected using customised outlier detection models. Detection of Meteor outbursts is difficult as they are generally highly aperiodic in nature and are thus difficult to study in detail. Further, it is likely that certain outbursts may be misclassified as periodic meteor showers. Thus, the detection, processing and extraction of meteor outbursts in the data was also extremely important for filtering the data. 3.2 Methods and Results The outliers were handled as follows. First, the ZHR values for every data point were calculated. The meteor count of the individual months across all the years were plotted against each month to observe the distribution of the meteor count as seen in figure 1. As meteors occur in cycles (Jenniskens, 2006), there would be a clear clustering of the points towards some modal value. Any outlier months were identified by using the score function in the Outliers (Komsta, 2011) package on R. The score function incorporated various distributions such as the normal, t, chi-squared, IQR and MAD. The outlier points that were identified corresponded to certain showers from a particular month and year. For each outlier point, all showers from that month and year were extracted. Each month and year in this list should contain a number of meteor outbursts which caused the meteor count to deviate from the modal value. For every month and year on this list, showers with a large deviation of the ZHR value were classified as meteor outbursts and were filtered out of the main data frame. When the meteor counts for each month was plotted again, there is a clear clustering of the points towards a modal value as seen in figure 1. 4 Feature Prediction 4.1 Introduction Feature Prediction involves predicting certain useful features of a meteor shower using machine learning models. The features that have been predicted are the Level, Shower Type and the Next Appearance of a meteor shower. The motivation behind predicting such features was to highlight the power of machine learning models in understanding trends in the meteor data. 4.2 Methods For each feature given below, we first detail the processing steps, then describe the machine learning model implemented along with the reasoning behind the decision. Figure 1 Distribution of Meteor Count Above : With Outbursts Below : Without Outbursts Comparison of the distribution of the meteor count for each month from 1990 to 2016. The monthly points noticeably clustered closer together clearly depict the removal of the meteor outbursts. 4.2.1 Level Prediction The Level of the meteor shower was calculated from the Number value which represented the number of meteors observed during the shower. The Level was defined as Low, Medium or High by the process of Equi-Depth binning of the Number feature. The following features were used for Level prediction : Start Month, Right ascension, Declination, Shower Type, Observatory and Solar Longitude. Only features that would be present before the predicted meteor shower were applied to simulate real time testing of the model. As Level prediction was a classification problem, we had to choose between the Decision Tree, K-Nearest Neighbours, Stochastic Gradient Descent, Support Vector Machines, Gaussian Process and Perceptron Classifiers. When a scoring algorithm was run on a small portion of the dataset, it was found that the Decision Tree Classifier and K-Nearest Neighbours had the least mean squared errors and were most suitable for the given data. Decision Tree Classifiers Decision Tree Classifiers were applied first to the data. The Decision Tree Classifier uses the training data to develop a Tree-like data structure and then classify the testing data based on this Tree. This model was ideal suited to the dataset because of its simplicity. The Decision Tree required simplified classification features and thus the Latitude, Longitude and Elevation features were combined into

4 WGN, The Journal of the IMO 45:5 (2017) the Observatory feature. Further, Shower types were also classified with an index to allow for simplicity in the Tree structure. K-Nearest Neighbour Classifiers Level prediction was then done with the use of a K-Nearest Neighbours Classifier. This classifier first learns the distribution of the data in the N-D space by dividing it into clusters which are centred around specific points, corresponding to the level value of the shower. It then classifies the testing data based on their relative distance to the centre of the various clusters. 4.2.2 Shower Type Prediction The Shower type data was present in the form of separate strings describing which type of shower it related to. By using the Label Encoder function from the Preprocessing module all the Shower types were indexed independently from 1 to 242 as there were 242 different meteor shower types in the data. The Label Encoder enabled us to later convert the indexes back to the Shower strings which could be then cross-referenced with the data. Only the Start Date was used for Shower type Prediction due to the shower type s repetitive nature. The Shower Type prediction was once again treated as a Classification problem and was scored with various classifier models (see section 4.2.1). When a scoring algorithm was run on a small portion of the dataset, the Decision Tree was found to give us the least mean squared error and was most suitable to the data. 4.2.3 Meteor Forecasting Meteor forecasting was done separately on the dataset consisting of the non-outbursts and the outbursts. This was implemented through an Long Short Term Memory (LSTM) network built on the Keras package in Python. The LSTM was advantageous as it kept track of one block of the data at a time thereby learning patterns within these blocks. This idea was specifically applicable to meteor forecasting as meteors occur in cycles which repeat periodically. By filtering the meteor outbursts out of the training set, the model could individually tackle the periodic meteor occurrences and then move on to the meteor outbursts dataset. The LSTM was used to predict the interval of time in hours between two successive meteor showers. 4.3 Analyses and Results It was found that the binning of the Shower feature and combination of the latitude, longitude and elevation into an observatory index greatly increased our prediction rates across all models. The individual analyses and prediction rates for each of the above features are described sequentially below. 4.3.1 Level Prediction It was seen that binning of the Number value played an important role in the predictive capacity of the model. By using values that aligned with layman descriptions of the level of the meteor shower, the model worked with exceedingly high accuracy. The data had an asymptotic distribution of the frequency of the Number data with the majority of the entries clustering at the lower Number values (see figure 2). Due to this distribution, it was natural for the model to deviate towards predicting at the lowest bin value repeatedly. This behaviour was checked by ensuring equal error rates for prediction of all of the individual bins. It was seen that both the Decision Tree Classifier and the K-Nearest Neighbour Classifier had equal error rates for all bins, ensuring the model was truly learning the trends in the data. Figure 2 Frequency Distribution of Number values Note the exponentially decreasing frequency as the meteor count value increases. This would cause the prediction model to stray towards predicting the lower meteor count values. Decision Tree Classifiers When implemented, the Decision Tree Classifier had a prediction rate at 80% for predicting the Level of the meteor shower against a random prediction rate of 33% (see table 1). Out of the 20% error rate, 35% of the errors were seen when the actual Number value of the meteor shower was within ±3 of the binning range, which corresponded to a close prediction. K-Nearest Neighbour Classifiers When implemented, the K-Nearest Neighbour Classifier had a prediction rate at 80% for predicting the Level of the shower against a random prediction rate of 33% (see table 1). Out of the 20% error rate, it was found that 37% of the errors were seen when the actual Number value of the meteor shower was within ±3 of the binning range which corresponded to a close prediction. 4.3.2 Shower Type Prediction The size of the Tree was customised as a large number of nodes had to be generated to account for the various

WGN, The Journal of the IMO 45:5 (2017) 5 Model Decision Tree K-Nearest Neighbour Prediction Rate Error for Bins 79.9% 1) 43.5% 2) 32.1% 3) 24.2% 80.5% 1) 55.6% 2) 23.6% 3) 20.7% Table 1 Level Prediction Values Numbering in the Error for Bins column corresponds to Bins 1 ([0, 5]), 2 ([5, 10]), 3 ([10, ]) types of meteor showers. Although the Decision Tree Classifier prediction model was based on predicting the exact shower type of the meteor shower(using no related showers), it was found that the prediction rate was at a high 41%, nearly a 100 better than a random prediction rate of 0.41%. To enable greater prediction, closely related showers must be binned into larger buckets and then the classifier must be run. It is highly possible that the large number of bins reduces our prediction percentage by a great extent. 4.3.3 Meteor Forecasting The LSTM was implemented on Keras with the motivation of predicting the hour wise time-gap between two consecutive meteor showers. The model had a prediction rate of 76%, with a loss factor of only 24% in the time-gap. The Root-Mean-Squared error for the dataset without outbursts was at 6.76 hours, which meant that the average error in prediction was 6.75 hours. For the outbursts, The Root-Mean-Squared-Error was at 9.53 hours. The distribution of the prediction of the time gap between consecutive meteor showers for both datasets are tabulated below (see tables 2 and 3). However, it must also be noticed that important factors like the comet orbit and the effect of Jupiter were not accounted for in the model. Range in Predictions hours within Range (-1, +1) 6.7% (-2, +2) 19.5% (-4, +4) 50.4% (-6, +6) 78.6% (-12, +12) 90.9% Table 2 Without Outbursts Forecasting Range in Predictions hours within Range (-1, +1) 0.7% (-2, +2) 1.9% (-4, +4) 5.8% (-6, +6) 16.5% (-12, +12) 76.3% 5 Conclusion In this paper, we have first analysed the IMO Visual Meteor Database to predict various features of meteor showers using Machine Learning models. We also detect meteor outbursts with the help of customised outlier detection algorithms. Using the Keras and Scikit-Learn modules on Python, supervised learning models were built and applied to the dataset for predicting the level, shower type and next occurrence of a meteor shower. This could potentially be used for making initial predictions of upcoming meteor showers. Acknowledgements This work was carried out at the Center for Fundamental Research and Creative Education (CFRCE), Bangalore, India under the guidance of Dr. B S Ramachandra whom I wish to acknowledge. I would like to acknowledge the Director Ms. Pratiti B R for creating the highly charged research atmosphere at CFRCE. Further I would like to thank Abhiram Harithas and Jyothish Jose, my mentors, for their invaluable guidance on this project. References Chollet F. et al. (2015). Keras : Machine learning in python. https://github.com/fchollet/keras. IMO (1982). Visual meteor database (vmdb). https://www.imo.net/members/imo vmdb. Jenniskens P. (2006). Meteor showers and their parent comets, chapter 1, page 7. Cambridge University Press. Komsta L. (2011). outliers: Tests for outliers. https://cran.r-project.org/package=outliers. R package version 0.14. McKinney W. (2010). Data structures for statistical computing in python. pages 51 56. Pedregosa F., Varoquaux G., Gramfort A., Michel V., Thirion B., Grisel O., Blondel M., Prettenhofer P., Weiss R., Dubourg V., Vanderplas J., Passos A., Cournapeau D., Brucher M., Perrot M., and douard Duchesnay (2011). Scikit-learn: Machine learning in python. Journal of Machine Learning Research, 12, 2825 2830. Vaubaillon J. (2017). A confidence index for forecasting of meteor showers. 143, 78 82. Veljkovic K. (2017). Metfns: Analysis of visual meteor data. https://cran.r-project.org/package=metfns. R package version 3.0.0. Table 3 Without Outbursts Forecasting