An Anomaly Detection Method for Spacecraft using Relevance Vector Learning
|
|
- Loreen Norris
- 5 years ago
- Views:
Transcription
1 An Anomaly Detection Method for Spacecraft using Relevance Vector Learning Ryohei Fujimaki 1, Takehisa Yairi 2, and Kazuo Machida 2 1 The Univ. of Tokyo, Aero. and Astronautics 2 The Univ. of Tokyo, RCAST Abstract. This paper proposes a novel anomaly detection system for spacecrafts based on data mining techniques. It constructs a nonlinear probabilistic model w.r.t. behavior of a spacecraft by applying the relevance vector regression and autoregression to massive telemetry data, and then monitors the on-line telemetry data using the model and detects anomalies. A major advantage over conventional anomaly detection methods is that this approach requires little a priori knowledge on the system. 1 Introduction Anomaly detection is a key issue in the development of recent advanced spacecraft. The space environment is very harsh for spacecraft due to a variety of factors such as direct radiation, great temperature difference, and so on. In addition, the space is so distant from the earth that it is extremely difficult to directly inspect and repair a damaged component. Therefore, early detection of anomalous symptoms is important to avoid disastrous situations such as loss of control. Although several anomaly detection/diagnosis methods using modern reasoning techniques have been developed, they have difficulties in acquiring accurate and complete models and knowledge of the spacecraft systems and in monitoring the system behavior exhaustively and efficiently. In this paper, we propose a new anomaly detection method for spacecraft based on data mining technique, autoregressive model and the relevance vector regression, and constructs a predictive model for each time series in the telemetry data. Then, it monitors online telemetry data and detect anomalies by checking the probability density of the observation. 2 Conventional Approaches to Anomaly Detection for Spacecraft Limit-checking checks whether the value is within the pre-defined upper and lower limits. Though the limit-checking can be applied to any types of spacecraft, it lacks flexibility and expressiveness and suffers from false alarm problem. In the model-based fault detection and diagnosis method, system models are utilized to simulate the spacecraft behavior and examine the validity of the
2 actual telemetry data. This approach would provide an ideal performance if an accurate and complete model and infinite computational power were available. In practice, however, both of them are not available. Expert systems also have been developed for this purpose. The knowledge is generally represented in the form of if-then production rules. Though the expert systems are powerful and flexible, it has a difficulty in preparing a set of accurate and complete knowledge on the spacecraft. In summary, the above methods have a common problem that they are too dependent on the knowledge of human experts. A reasonable approach to this problem is the application of data mining and machine learning techniques to the telemetry data. Actually, some researchers have developed anomaly detection methods for spacecraft using regression tree learning[7], temporal pattern clustering[8], association rule mining[9]. 3 Proposed Anomaly Detection System 3.1 Autoregressive Model Autoregressive (AR) model is the most basic data mining technique for timeseries data[3][10][11]. For the purpose of applying AR model to anomaly detection problems for spacecraft, we define AR model as τ j ARk = Θj x k (1) where j = {1,, s} represents the jth series of telemetry data, Θ j = (Θ j 0, Θj 1,k 1,..., Θ j 1,k p,..., Θj s,k 1,..., Θj s,k p ) is a row vector of AR coefficients, x k = (1, τk 1 1,..., τ k p 1,..., τ k 1 s,..., τ k p s ) is the data vector of all p previous time series, and the notation AR represent that the target value τ is based on AR model. This modified AR model implies that the value of a target time-series depends not only on the past values of itself but also on those of other series. The capability of modeling the relationships among some series is a great advantage. Fig.1 is the concept of this model. We made use of the framework of the relevance vector learning to extend this model to nonlinear and probabilistic. Fig. 1. Concept of our AR model.
3 3.2 Relevance Vector Regression The relevance vector regression (RVR) originally proposed by Tipping[5] is a state of the art kernel-based nonlinear regression learning method[1][2][4][6]. We write a sample of N training pairs as {x n, t n } N n=1 for the jth telemetry series, corresponding x k and τ j k. Hereinafter, we deal with the jth series of telemetry data and omit the notation j for simplification. The RVR model assumes that the targets are samples from a distribution model with additive independent zero-mean Gaussian noise, with variance σ 2, t = y + ɛ = Φw + ɛ (2) where t = (t 1,..., t N ), y, Φ = (φ 1,..., φ M ), w = (w 0,..., w M ) T, φ m (x) = K(x, x m ) and ɛ = (ɛ 1,..., ɛ n ) T represent the target vector, the vector of the predicted value, the N M design matrix, the weights, the kernel function, and an error vector, respectively. To achieve the sparsity, M independent hyperparameters, α = (α 1,..., α M ) T over w are indroduced. Then, after maximizing logarithm likelihood w.r.t. α and σ 2, we obtain a conditional distribution model for a new datum x as p(t t, α MP, σ 2 MP ) = N (t y, σ 2 ) (3) where y is the predict value of the new target t and σ 2 is the variance of the prediction. See [1][4][6] for more details of the sparse Bayesian learning. 3.3 Anomaly Detection System The anomaly detection system based on the proposed method operates as follows, 1. (Learning) Learn the relevance vector autoregressive model using a set of validated normal telemetry data. 2. (Prediction) Compute the next probable range of the target series. 3. (Monitoring) Obtain and check the (pseudo-)telemetry data. 4. (Alarming) Give an alarm if the data is out of the predicted range. 5. Repeat steps 2-4 The system is supposed to give many false alarms if we directly apply the redefined AR model Eq.(1) and the RVR Eq.(3). The reasons are, 1. The AR model completely cannot be modeled as the complicate spacecraft system, t (x ) = t AR (x ) = t T RUE (x ) + ɛ AR = y (x ) + ɛ (x ) p(ɛ ) = N (0, σ 2 ) (4) where ɛ AR is the modeling error of the AR model. 2. The RVR adopts the relevance vector for the prototypes of the model formed as Eq.(2) 3. 3 This is a great advantage of the RVR for the execution speed of anomaly detection.
4 As the result, some training data can be mistaken as anomaly though all training data are normal because these data have relatively so large ɛ AR that the system regards them as the data which is far from the prototype data. We evaluated the variance of the difference between another data set {x i, t i } N 2 i=1 and corresponding prediction values, y (x i ). ˆσ AR 2 = 1 N 2 (t i y (x i )) 2, N 2 i N 2 i ɛ(x i ) 0 (5) If the AR model completely describes the system behavior, (5) must be zero in theory. Therefore, we extended (3) as p(t t, α MP, σ 2 MP ) = N (y, σ 2 + ˆσ 2 AR). (6) 4 Experiment and Discussion We performed an experiment with telemetry data obtained from an orbital rendezvous simulation. This telemetry consists of 27 time-series variables in total, 13 of which are from position and attitude control subsystem, and the rest are from propulsion subsystem. In more detail, the former group consists of 12 numerical observation time-series variables regarding the position and attitude of the vehicle and one command sequence. The latter group consists of 14 discretevalued time-series variables, each of which indicates the command input to each of the 14 thruster engines. In this experiment, we assumed a scenario where the power of fourth thruster engine used for the pitch control falls to zero at time 250 [sec]. With this scenario, we performed following two scenarios. Comparison of Proposed Method with normal RV Autoregressive Model First, we have compared the proposed method with the normal relevance vector autoregressive model. Fig.2, Fig.3 show the results of anomaly detection in the series which represents the pitch angle. The solid line in the upper figure shows the pitch angle and the dotted line shows the predicted range, and the solid line in the lower figure shows the probability density of the observation. The system gives the alarm when the probability became lower than the computed limit 4. We can see that the normal model gives many false alarms. On the other hand, the proposed method correctly gives alarms after a little time delay. The proposed method also succeeded in detecting anomalies in the series representing the pitch rate as shown in Fig.4. Comparison with Conventional Limit-checking We compared the proposed method with the conventional limit-checking. We 4 In this experiment, we adopted the probability density value of σ 2 + σar 2 as limit.
5 Fig. 2. Result of anomaly detection by normal RV AR model. The upper graph shows the confidence of target series in the telemetry data. Fig. 3. Result of anomaly detection by the proposed method. The lower graph shows the probability density. set the limit on the standard deviation for the proposed method and on the maximum absolute value in the normal phase for the limit-checking. Fig.4 shows the result. The conventional limit-checking fails to detect slight anomalies like this case. On the other hand, the proposed method is capable of detecting this anomaly because it can dynamically estimate proper range of the target series. Fig. 4. Comparison of the proposed method with conventional limit-checking. The solid line represents the pitch rate and the dotted lines represent the upper and lower bounds given by each method. We only showed the results with respect to the series which included the anomalies due to limitations of space, though we have ran the experiments against all series. 5 Conclusion This paper proposed a new anomaly detection method based on the relevance vector regression and autoregressive model. First, we extended the traditional AR model (3.1) and adopted the relevance vector frameworks for learning this model. In addition, we extended this RV AR model for the purpose of removing false alarm (3.3).
6 Compared with the conventional anomaly detection method, this method has great advantages. First, the proposed method requires little a priori knowledge on the spacecraft system. Therefore, it can be applied to various kinds of spacecraft. We performed an experiment with telemetry data obtained from an orbital rendezvous simulation and confirmed the efficiency of the propos4ed method. Acknowledgements The authors would like to thank Japan Aerospace Exploration Agency (JAXA) for providing the simulation telemetry data. References 1. A.C. Faul and M.E. Tipping. Analysis of Sparse Bayesian Learning, In T.G. Dietterich, S. Becker and Z. Ghahramanim editors, Advances in Neural Information Processing Systems 14, pages , MIT Press, C.M. Bishop and M.E. Tipping, Variational Relevance Vector Machines, In C. Boutilier and M. Goldszmidt, editors, Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence, page 46-53, Morgan Kaufmann, K.-R. Muller, A.J. Smola, G.Ratsch, B. Scholkopf, J. Kohlmorgen, and V. Vapnik, Predicting time series with support vector machines. In W. Gerstner, A. Germond, M. Hasler, and J.-D. Nicoud, editors, Artificail Neural Networks, ICANN 97, page , Berlin, 1997, Springer Lecture Notes in Computer Science, Vol M.E. Tipping, Sparse Bayesian Learning and the Relevance Vector Machinem Journal of Machine Learning Researchm 1: , M.E. Tipping The Relevance Vector Machine. In S.A. Solla, T.K. Leen, and K.- R. Muller, editors, Advances in Neural Information Processing Systems 12, pages MIT Press M.E. Tipping, and A. C. Faul (2003). Fast marginal likelihood maximisation for sparse Bayesian models. In C. M. Bishop and B. J. Frey (Eds.), Proceedings of the Ninth International Workshop on Artificial Intelligence and Statistics, Key West, FL, Jan Minoru Nakatsugawa, Takehisa Yairi, Naoki Ishihama, Koichi Hori and Shinichi Nakasuka Supporting Anomaly Detection from Satellite Telemetry Data by Regression Trees The 24th International Symposium on Space Technology and Science (ISTS), Takehisa Yairi, Shiro Ogasawara, Koichi Hori, Shinichi Nakasuka, and Naoki Ishihama Summarization of Spacecrafts Telemetry Data By Extracting Significant Temporal Patterns The Eighth Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD2004), pp , Takehisa Yairi, Yoshikiyo Kato and Koichi Hori, Fault Detection by Mining Association Rules from House-keeping Data, Proc. of International Symposium on Artificial Intelligence, Robotics and Automation in Space (i-sairas 2001) 10. U.U. Muller, A. Schick and W. Wefelmeyer(2004), Efficient prediction for linear and nonlinear autoregressive models, 2004, (Submitted paper) 11. W.D. Penny and S.J. Roberts, Bayesian Methods for Autoregressive Models, IEEE Workshop on Neural Networks for Signal Processing, Sydney Australia, December 2000.
ADAPTIVE LIMIT-CHECKING FOR SPACECRAFT USING RELEVANCE VECTOR AUTOREGRESSIVE MODEL. 2 RCAST, Univ.
ADAPTIVE LIMIT-CHECKING FOR SPACECRAFT USING RELEVANCE VECTOR AUTOREGRESSIVE MODEL Ryohei Fujimaki 1, Takehisa Yairi 2, and Kazuo Machida 2 1 Dept. of Aero. and Astronautics, Univ. of Tokyo, 4-6-1 Komaba
More informationFast Marginal Likelihood Maximisation for Sparse Bayesian Models
Fast Marginal Likelihood Maximisation for Sparse Bayesian Models Michael E. Tipping and Anita C. Faul Microsoft Research, Cambridge, U.K.....................................................................
More informationMaximum Direction to Geometric Mean Spectral Response Ratios using the Relevance Vector Machine
Maximum Direction to Geometric Mean Spectral Response Ratios using the Relevance Vector Machine Y. Dak Hazirbaba, J. Tezcan, Q. Cheng Southern Illinois University Carbondale, IL, USA SUMMARY: The 2009
More informationESANN'2001 proceedings - European Symposium on Artificial Neural Networks Bruges (Belgium), April 2001, D-Facto public., ISBN ,
Sparse Kernel Canonical Correlation Analysis Lili Tan and Colin Fyfe 2, Λ. Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong. 2. School of Information and Communication
More informationNON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION. Haiqin Yang, Irwin King and Laiwan Chan
In The Proceedings of ICONIP 2002, Singapore, 2002. NON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION Haiqin Yang, Irwin King and Laiwan Chan Department
More informationSmooth Bayesian Kernel Machines
Smooth Bayesian Kernel Machines Rutger W. ter Borg 1 and Léon J.M. Rothkrantz 2 1 Nuon NV, Applied Research & Technology Spaklerweg 20, 1096 BA Amsterdam, the Netherlands rutger@terborg.net 2 Delft University
More informationProbabilistic Regression Using Basis Function Models
Probabilistic Regression Using Basis Function Models Gregory Z. Grudic Department of Computer Science University of Colorado, Boulder grudic@cs.colorado.edu Abstract Our goal is to accurately estimate
More informationRelevance Vector Machines for Earthquake Response Spectra
2012 2011 American American Transactions Transactions on on Engineering Engineering & Applied Applied Sciences Sciences. American Transactions on Engineering & Applied Sciences http://tuengr.com/ateas
More informationUsing Gaussian Processes for Variance Reduction in Policy Gradient Algorithms *
Proceedings of the 8 th International Conference on Applied Informatics Eger, Hungary, January 27 30, 2010. Vol. 1. pp. 87 94. Using Gaussian Processes for Variance Reduction in Policy Gradient Algorithms
More informationLinear Dependency Between and the Input Noise in -Support Vector Regression
544 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 14, NO. 3, MAY 2003 Linear Dependency Between the Input Noise in -Support Vector Regression James T. Kwok Ivor W. Tsang Abstract In using the -support vector
More informationDiscriminative Direction for Kernel Classifiers
Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering
More informationLearning Gaussian Process Models from Uncertain Data
Learning Gaussian Process Models from Uncertain Data Patrick Dallaire, Camille Besse, and Brahim Chaib-draa DAMAS Laboratory, Computer Science & Software Engineering Department, Laval University, Canada
More informationNeutron inverse kinetics via Gaussian Processes
Neutron inverse kinetics via Gaussian Processes P. Picca Politecnico di Torino, Torino, Italy R. Furfaro University of Arizona, Tucson, Arizona Outline Introduction Review of inverse kinetics techniques
More informationRelevance Vector Machines
LUT February 21, 2011 Support Vector Machines Model / Regression Marginal Likelihood Regression Relevance vector machines Exercise Support Vector Machines The relevance vector machine (RVM) is a bayesian
More informationPILCO: A Model-Based and Data-Efficient Approach to Policy Search
PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol
More informationVariational Principal Components
Variational Principal Components Christopher M. Bishop Microsoft Research 7 J. J. Thomson Avenue, Cambridge, CB3 0FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop In Proceedings
More informationRecent Advances in Bayesian Inference Techniques
Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian
More informationSupport Vector Machine Regression for Volatile Stock Market Prediction
Support Vector Machine Regression for Volatile Stock Market Prediction Haiqin Yang, Laiwan Chan, and Irwin King Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 2: Bayesian Basics https://people.orie.cornell.edu/andrew/orie6741 Cornell University August 25, 2016 1 / 17 Canonical Machine Learning
More information1162 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 5, SEPTEMBER The Evidence Framework Applied to Support Vector Machines
1162 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 5, SEPTEMBER 2000 Brief Papers The Evidence Framework Applied to Support Vector Machines James Tin-Yau Kwok Abstract In this paper, we show that
More informationGWAS IV: Bayesian linear (variance component) models
GWAS IV: Bayesian linear (variance component) models Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS IV: Bayesian
More informationBayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine
Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Mike Tipping Gaussian prior Marginal prior: single α Independent α Cambridge, UK Lecture 3: Overview
More information2 Related Works. 1 Introduction. 2.1 Gaussian Process Regression
Gaussian Process Regression with Dynamic Active Set and Its Application to Anomaly Detection Toshikazu Wada 1, Yuki Matsumura 1, Shunji Maeda 2, and Hisae Shibuya 3 1 Faculty of Systems Engineering, Wakayama
More informationSupport Vector Regression with Automatic Accuracy Control B. Scholkopf y, P. Bartlett, A. Smola y,r.williamson FEIT/RSISE, Australian National University, Canberra, Australia y GMD FIRST, Rudower Chaussee
More informationA CUSUM approach for online change-point detection on curve sequences
ESANN 22 proceedings, European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning. Bruges Belgium, 25-27 April 22, i6doc.com publ., ISBN 978-2-8749-49-. Available
More informationVirtual Sensors and Large-Scale Gaussian Processes
Virtual Sensors and Large-Scale Gaussian Processes Ashok N. Srivastava, Ph.D. Principal Investigator, IVHM Project Group Lead, Intelligent Data Understanding ashok.n.srivastava@nasa.gov Coauthors: Kamalika
More informationDynamical Modeling with Kernels for Nonlinear Time Series Prediction
Dynamical Modeling with Kernels for Nonlinear Time Series Prediction Liva Ralaivola Laboratoire d Informatique de Paris 6 Université Pierre et Marie Curie 8, rue du capitaine Scott F-75015 Paris, FRANCE
More informationReliability Monitoring Using Log Gaussian Process Regression
COPYRIGHT 013, M. Modarres Reliability Monitoring Using Log Gaussian Process Regression Martin Wayne Mohammad Modarres PSA 013 Center for Risk and Reliability University of Maryland Department of Mechanical
More informationSupport Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM
1 Support Vector Machines (SVM) in bioinformatics Day 1: Introduction to SVM Jean-Philippe Vert Bioinformatics Center, Kyoto University, Japan Jean-Philippe.Vert@mines.org Human Genome Center, University
More informationMissing Data in Kernel PCA
Missing Data in Kernel PCA Guido Sanguinetti, Neil D. Lawrence Department of Computer Science, University of Sheffield 211 Portobello Street, Sheffield S1 4DP, U.K. 19th June 26 Abstract Kernel Principal
More informationNovelty Detection based on Extensions of GMMs for Industrial Gas Turbines
Novelty Detection based on Extensions of GMMs for Industrial Gas Turbines Yu Zhang, Chris Bingham, Michael Gallimore School of Engineering University of Lincoln Lincoln, U.. {yzhang; cbingham; mgallimore}@lincoln.ac.uk
More informationDeep Gaussian Processes
Deep Gaussian Processes Neil D. Lawrence 30th April 2015 KTH Royal Institute of Technology Outline Introduction Deep Gaussian Process Models Variational Approximation Samples and Results Outline Introduction
More informationLecture 3. Linear Regression II Bastian Leibe RWTH Aachen
Advanced Machine Learning Lecture 3 Linear Regression II 02.11.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de This Lecture: Advanced Machine Learning Regression
More informationGaussian with mean ( µ ) and standard deviation ( σ)
Slide from Pieter Abbeel Gaussian with mean ( µ ) and standard deviation ( σ) 10/6/16 CSE-571: Robotics X ~ N( µ, σ ) Y ~ N( aµ + b, a σ ) Y = ax + b + + + + 1 1 1 1 1 1 1 1 1 1, ~ ) ( ) ( ), ( ~ ), (
More informationAdaptive Sparseness Using Jeffreys Prior
Adaptive Sparseness Using Jeffreys Prior Mário A. T. Figueiredo Institute of Telecommunications and Department of Electrical and Computer Engineering. Instituto Superior Técnico 1049-001 Lisboa Portugal
More informationLearning Non-stationary System Dynamics Online using Gaussian Processes
Learning Non-stationary System Dynamics Online using Gaussian Processes Axel Rottmann and Wolfram Burgard Department of Computer Science, University of Freiburg, Germany Abstract. Gaussian processes are
More informationDynamical Modeling with Kernels for Nonlinear Time Series Prediction
Dynamical Modeling with Kernels for Nonlinear Time Series Prediction Liva Ralaivola Laboratoire d Informatique de Paris 6 Université Pierre et Marie Curie 8, rue du capitaine Scott F-75015 Paris, FRANCE
More informationKeywords: Multimode process monitoring, Joint probability, Weighted probabilistic PCA, Coefficient of variation.
2016 International Conference on rtificial Intelligence: Techniques and pplications (IT 2016) ISBN: 978-1-60595-389-2 Joint Probability Density and Weighted Probabilistic PC Based on Coefficient of Variation
More informationMultiple-step Time Series Forecasting with Sparse Gaussian Processes
Multiple-step Time Series Forecasting with Sparse Gaussian Processes Perry Groot ab Peter Lucas a Paul van den Bosch b a Radboud University, Model-Based Systems Development, Heyendaalseweg 135, 6525 AJ
More informationNonparameteric Regression:
Nonparameteric Regression: Nadaraya-Watson Kernel Regression & Gaussian Process Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationKernel methods and the exponential family
Kernel methods and the exponential family Stéphane Canu 1 and Alex J. Smola 2 1- PSI - FRE CNRS 2645 INSA de Rouen, France St Etienne du Rouvray, France Stephane.Canu@insa-rouen.fr 2- Statistical Machine
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationSupport Vector Machines
Support Vector Machines Tobias Pohlen Selected Topics in Human Language Technology and Pattern Recognition February 10, 2014 Human Language Technology and Pattern Recognition Lehrstuhl für Informatik 6
More informationMachine learning for pervasive systems Classification in high-dimensional spaces
Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationApproximate Kernel PCA with Random Features
Approximate Kernel PCA with Random Features (Computational vs. Statistical Tradeoff) Bharath K. Sriperumbudur Department of Statistics, Pennsylvania State University Journées de Statistique Paris May 28,
More informationValue Function Approximation through Sparse Bayesian Modeling
Value Function Approximation through Sparse Bayesian Modeling Nikolaos Tziortziotis and Konstantinos Blekas Department of Computer Science, University of Ioannina, P.O. Box 1186, 45110 Ioannina, Greece,
More informationStatistical Techniques in Robotics (16-831, F12) Lecture#21 (Monday November 12) Gaussian Processes
Statistical Techniques in Robotics (16-831, F12) Lecture#21 (Monday November 12) Gaussian Processes Lecturer: Drew Bagnell Scribe: Venkatraman Narayanan 1, M. Koval and P. Parashar 1 Applications of Gaussian
More informationOutliers Treatment in Support Vector Regression for Financial Time Series Prediction
Outliers Treatment in Support Vector Regression for Financial Time Series Prediction Haiqin Yang, Kaizhu Huang, Laiwan Chan, Irwin King, and Michael R. Lyu Department of Computer Science and Engineering
More informationDistinguishing Causes from Effects using Nonlinear Acyclic Causal Models
JMLR Workshop and Conference Proceedings 6:17 164 NIPS 28 workshop on causality Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models Kun Zhang Dept of Computer Science and HIIT University
More informationProbabilistic Graphical Models for Image Analysis - Lecture 1
Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.
More informationModeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop
Modeling Data with Linear Combinations of Basis Functions Read Chapter 3 in the text by Bishop A Type of Supervised Learning Problem We want to model data (x 1, t 1 ),..., (x N, t N ), where x i is a vector
More informationIntroduction to SVM and RVM
Introduction to SVM and RVM Machine Learning Seminar HUS HVL UIB Yushu Li, UIB Overview Support vector machine SVM First introduced by Vapnik, et al. 1992 Several literature and wide applications Relevance
More informationCOMP 551 Applied Machine Learning Lecture 20: Gaussian processes
COMP 55 Applied Machine Learning Lecture 2: Gaussian processes Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~hvanho2/comp55
More informationIntelligent Systems I
Intelligent Systems I 00 INTRODUCTION Stefan Harmeling & Philipp Hennig 24. October 2013 Max Planck Institute for Intelligent Systems Dptmt. of Empirical Inference Which Card? Opening Experiment Which
More informationPATTERN RECOGNITION AND MACHINE LEARNING
PATTERN RECOGNITION AND MACHINE LEARNING Chapter 1. Introduction Shuai Huang April 21, 2014 Outline 1 What is Machine Learning? 2 Curve Fitting 3 Probability Theory 4 Model Selection 5 The curse of dimensionality
More informationLearning Non-stationary System Dynamics Online Using Gaussian Processes
Learning Non-stationary System Dynamics Online Using Gaussian Processes Axel Rottmann and Wolfram Burgard Department of Computer Science, University of Freiburg, Germany Abstract. Gaussian processes are
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationPattern Recognition and Machine Learning. Bishop Chapter 6: Kernel Methods
Pattern Recognition and Machine Learning Chapter 6: Kernel Methods Vasil Khalidov Alex Kläser December 13, 2007 Training Data: Keep or Discard? Parametric methods (linear/nonlinear) so far: learn parameter
More informationLeast Absolute Shrinkage is Equivalent to Quadratic Penalization
Least Absolute Shrinkage is Equivalent to Quadratic Penalization Yves Grandvalet Heudiasyc, UMR CNRS 6599, Université de Technologie de Compiègne, BP 20.529, 60205 Compiègne Cedex, France Yves.Grandvalet@hds.utc.fr
More informationbelow, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing
Kernel PCA Pattern Reconstruction via Approximate Pre-Images Bernhard Scholkopf, Sebastian Mika, Alex Smola, Gunnar Ratsch, & Klaus-Robert Muller GMD FIRST, Rudower Chaussee 5, 12489 Berlin, Germany fbs,
More informationMultivariate statistical methods and data mining in particle physics
Multivariate statistical methods and data mining in particle physics RHUL Physics www.pp.rhul.ac.uk/~cowan Academic Training Lectures CERN 16 19 June, 2008 1 Outline Statement of the problem Some general
More informationReading Group on Deep Learning Session 1
Reading Group on Deep Learning Session 1 Stephane Lathuiliere & Pablo Mesejo 2 June 2016 1/31 Contents Introduction to Artificial Neural Networks to understand, and to be able to efficiently use, the popular
More informationMachine Learning. 7. Logistic and Linear Regression
Sapienza University of Rome, Italy - Machine Learning (27/28) University of Rome La Sapienza Master in Artificial Intelligence and Robotics Machine Learning 7. Logistic and Linear Regression Luca Iocchi,
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationLearning Binary Classifiers for Multi-Class Problem
Research Memorandum No. 1010 September 28, 2006 Learning Binary Classifiers for Multi-Class Problem Shiro Ikeda The Institute of Statistical Mathematics 4-6-7 Minami-Azabu, Minato-ku, Tokyo, 106-8569,
More informationUsing Kernel PCA for Initialisation of Variational Bayesian Nonlinear Blind Source Separation Method
Using Kernel PCA for Initialisation of Variational Bayesian Nonlinear Blind Source Separation Method Antti Honkela 1, Stefan Harmeling 2, Leo Lundqvist 1, and Harri Valpola 1 1 Helsinki University of Technology,
More informationA Bayesian Committee Machine
Neural Computation, Vol. 12, pages 2719-2741, 2000 A Bayesian Committee Machine Volker Tresp Siemens AG, Corporate Technology Dept. Information and Communications Otto-Hahn-Ring 6, 81730 Munich, Germany
More informationSVM Incremental Learning, Adaptation and Optimization
SVM Incremental Learning, Adaptation and Optimization Christopher P. Diehl Applied Physics Laboratory Johns Hopkins University Laurel, MD 20723 Chris.Diehl@jhuapl.edu Gert Cauwenberghs ECE Department Johns
More informationLogistic Regression and Boosting for Labeled Bags of Instances
Logistic Regression and Boosting for Labeled Bags of Instances Xin Xu and Eibe Frank Department of Computer Science University of Waikato Hamilton, New Zealand {xx5, eibe}@cs.waikato.ac.nz Abstract. In
More informationSTA 4273H: Sta-s-cal Machine Learning
STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our
More informationLatent Variable Models with Gaussian Processes
Latent Variable Models with Gaussian Processes Neil D. Lawrence GP Master Class 6th February 2017 Outline Motivating Example Linear Dimensionality Reduction Non-linear Dimensionality Reduction Outline
More informationKnowledge-Based Nonlinear Kernel Classifiers
Knowledge-Based Nonlinear Kernel Classifiers Glenn M. Fung, Olvi L. Mangasarian, and Jude W. Shavlik Computer Sciences Department, University of Wisconsin Madison, WI 5376 {gfung,olvi,shavlik}@cs.wisc.edu
More informationCSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13
CSE 291. Assignment 3 Out: Wed May 23 Due: Wed Jun 13 3.1 Spectral clustering versus k-means Download the rings data set for this problem from the course web site. The data is stored in MATLAB format as
More informationCOMP 551 Applied Machine Learning Lecture 21: Bayesian optimisation
COMP 55 Applied Machine Learning Lecture 2: Bayesian optimisation Associate Instructor: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp55 Unless otherwise noted, all material posted
More informationUncertainties estimation and compensation on a robotic manipulator. Application on Barrett arm
Uncertainties estimation and compensation on a robotic manipulator. Application on Barrett arm Duong Dang December, 8 Abstract In this project, two different regression methods have been used to estimate
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 218 Outlines Overview Introduction Linear Algebra Probability Linear Regression 1
More informationBayesian methods in economics and finance
1/26 Bayesian methods in economics and finance Linear regression: Bayesian model selection and sparsity priors Linear Regression 2/26 Linear regression Model for relationship between (several) independent
More informationSupport Vector Machines
Support Vector Machines Stephan Dreiseitl University of Applied Sciences Upper Austria at Hagenberg Harvard-MIT Division of Health Sciences and Technology HST.951J: Medical Decision Support Overview Motivation
More informationPrediction of double gene knockout measurements
Prediction of double gene knockout measurements Sofia Kyriazopoulou-Panagiotopoulou sofiakp@stanford.edu December 12, 2008 Abstract One way to get an insight into the potential interaction between a pair
More informationABSTRACT INTRODUCTION
ABSTRACT Presented in this paper is an approach to fault diagnosis based on a unifying review of linear Gaussian models. The unifying review draws together different algorithms such as PCA, factor analysis,
More informationMatching the dimensionality of maps with that of the data
Matching the dimensionality of maps with that of the data COLIN FYFE Applied Computational Intelligence Research Unit, The University of Paisley, Paisley, PA 2BE SCOTLAND. Abstract Topographic maps are
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V
More informationDeep Gaussian Processes
Deep Gaussian Processes Neil D. Lawrence 8th April 2015 Mascot Num 2015 Outline Introduction Deep Gaussian Process Models Variational Methods Composition of GPs Results Outline Introduction Deep Gaussian
More informationForecast daily indices of solar activity, F10.7, using support vector regression method
Research in Astron. Astrophys. 9 Vol. 9 No. 6, 694 702 http://www.raa-journal.org http://www.iop.org/journals/raa Research in Astronomy and Astrophysics Forecast daily indices of solar activity, F10.7,
More informationApproximate Kernel Methods
Lecture 3 Approximate Kernel Methods Bharath K. Sriperumbudur Department of Statistics, Pennsylvania State University Machine Learning Summer School Tübingen, 207 Outline Motivating example Ridge regression
More informationEEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1
EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle
More informationLearning features by contrasting natural images with noise
Learning features by contrasting natural images with noise Michael Gutmann 1 and Aapo Hyvärinen 12 1 Dept. of Computer Science and HIIT, University of Helsinki, P.O. Box 68, FIN-00014 University of Helsinki,
More informationA GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong
A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES Wei Chu, S. Sathiya Keerthi, Chong Jin Ong Control Division, Department of Mechanical Engineering, National University of Singapore 0 Kent Ridge Crescent,
More informationAnalysis of Multiclass Support Vector Machines
Analysis of Multiclass Support Vector Machines Shigeo Abe Graduate School of Science and Technology Kobe University Kobe, Japan abe@eedept.kobe-u.ac.jp Abstract Since support vector machines for pattern
More informationGaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008
Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:
More informationGaussian Processes (10/16/13)
STA561: Probabilistic machine learning Gaussian Processes (10/16/13) Lecturer: Barbara Engelhardt Scribes: Changwei Hu, Di Jin, Mengdi Wang 1 Introduction In supervised learning, we observe some inputs
More informationOn the Noise Model of Support Vector Machine Regression. Massimiliano Pontil, Sayan Mukherjee, Federico Girosi
MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL AND COMPUTATIONAL LEARNING DEPARTMENT OF BRAIN AND COGNITIVE SCIENCES A.I. Memo No. 1651 October 1998
More informationGaussian processes and bayesian optimization Stanisław Jastrzębski. kudkudak.github.io kudkudak
Gaussian processes and bayesian optimization Stanisław Jastrzębski kudkudak.github.io kudkudak Plan Goal: talk about modern hyperparameter optimization algorithms Bayes reminder: equivalent linear regression
More informationOn-line Support Vector Machines and Optimization Strategies
On-line Support Vector Machines and Optimization Strategies J.M. Górriz, C.G. Puntonet, M. Salmerón, R. Martin-Clemente and S. Hornillo-Mellado and J. Ortega E.S.I., Informática, Universidad de Granada
More informationESANN'2003 proceedings - European Symposium on Artificial Neural Networks Bruges (Belgium), April 2003, d-side publi., ISBN X, pp.
On different ensembles of kernel machines Michiko Yamana, Hiroyuki Nakahara, Massimiliano Pontil, and Shun-ichi Amari Λ Abstract. We study some ensembles of kernel machines. Each machine is first trained
More informationAn Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM
An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM Wei Chu Chong Jin Ong chuwei@gatsby.ucl.ac.uk mpeongcj@nus.edu.sg S. Sathiya Keerthi mpessk@nus.edu.sg Control Division, Department
More informationIntroduction Dual Representations Kernel Design RBF Linear Reg. GP Regression GP Classification Summary. Kernel Methods. Henrik I Christensen
Kernel Methods Henrik I Christensen Robotics & Intelligent Machines @ GT Georgia Institute of Technology, Atlanta, GA 30332-0280 hic@cc.gatech.edu Henrik I Christensen (RIM@GT) Kernel Methods 1 / 37 Outline
More informationarxiv: v1 [stat.ml] 10 Dec 2017
Sensitivity Analysis for Predictive Uncertainty in Bayesian Neural Networks Stefan Depeweg 1,2, José Miguel Hernández-Lobato 3, Steffen Udluft 2, Thomas Runkler 1,2 arxiv:1712.03605v1 [stat.ml] 10 Dec
More information