CONSTRUCTING A NOVEL MONOTONICITY CONSTRAINED SUPPORT VECTOR REGRESSION MODEL
|
|
- Gabriel Robertson
- 5 years ago
- Views:
Transcription
1 CONSTRUCTING A NOVEL MONOTONICITY CONSTRAINED SUPPORT VECTOR REGRESSION MODEL Chih-Chuan Chen, Department of Industrial and Information management, National Cheng Kung University, Taiwan, R.O.C. Department of Leisure Information Management Taiwan Shoufu University, Taiwan, R.O.C. ccchen@tsu.edu.tw Shu-Ching Kuo, Department of Leisure Information Management Taiwan Shoufu University, Taiwan, R.O.C. su0102@gmail.com Corresponding Author Sheng-Tun Li, Department of Industrial and Information management National Cheng Kung University, Taiwan, R.O.C. Institute of Information Management National Cheng Kung University, Taiwan, R.O.C. stli@mail.ncku.edu.tw ABSTRACT This paper aims to construct a monotonicity constrained nonlinear regression model based on Support Vector Machines (SVMs). In many application areas of machine learning, there exists prior knowledge concerning the monotone relations between the response variable and some of the predictor variables. Monotonicity may be an important model requirement with a view toward explaining and justifying decisions. Therefore, in the study we propose a monotonicity constrained Support Vector Regression (SVR) model that incorporates in the monotone nature of the problems. A quadratic programming problem in the dual space is developed similar to its SVR predecessor. When applied to some synthetic data sets, the proposed method shows advantages and promising results. Keywords: Classification problems, SVM, Monotonicity constraints S5-241
2 INTRODUCTION Data mining techniques enable us to discover hidden patterns and extract valuable knowledge from databases. With the advent of the computer, various data mining methods have been proposed and vehemently discussed. Among them, support vector machines (SVM), characterized by convex optimization problems, is an important method in the fields of neural networks and nonlinear modeling, and has been successfully applied to problems of classification and nonlinear function estimation. The technology of SVM, pioneered by Vapnik in 1995, is a state-of-the-art artificial neural network (ANN) based on statistical learning (Vapnik, 1995; Vapnik, 1998). In recent years, it has drawn overwhelming attention from diverse research communities due to its outstanding performance to solve classification problems and its novel approach to improve the generalization property of ANNs (Burges, 1998; Cristianini & Shawe-Taylor, 2000). Unlike ANNs which minimize empirical risk, SVM is designed to minimize the structural risk by minimizing an upper bound of the generalization error rather than the training error. Therefore, the over fitting problem in machine learning can be solved successfully. Compared to ANNs, the other outstanding property of SVM is that the task of training SVM can be mapped to a uniquely solvable linearly constrained quadratic programming problem, which produces a solution that is always unique and globally optimal. SVMs have been widely applied to many kinds of fields in the past few years, such as corporate distress, consumer loan evaluation, text categorization, handwritten digit recognition, speaker verification, bioinformatics, and many others. In many applications of classification, we have a priori knowledge to the extent that, all else being equal, an increase in an input variable should not lead to a decrease (or increase) in class label. For example, if loan applicants A and B have the same attribute values, except that A has a higher income than B, then it would be surprising if B got the loan while A did not. Examples of other application domains in which we can have this type of knowledge are legal support systems, medicine (e.g. smoking increases the probability of vascular diseases), operations research and economics (e.g. house prices increase with the house area). In the aforementioned problems, one can see that there are some monotonic relationships between the class and some of the attributes. When taking into account this prior knowledge about the data, one needs to add some monotonicity constraints into the classification model like SVM. It has been shown that classification technique incorporated with monotonicity constraints can extract knowledge with more justifiability and comprehensibility. In the data-mining literatures about monotonicity constraints, there are two different approaches for dealing with problems that have prior knowledge of monotonic properties, although there are only few papers focused on this topic. One is to apply a relabeling technique to those data missing monotonicity (Duivesteijn & Feelders, 2008). The other is to add the monotonicity constraints directly to the optimization modeling settings (Falck et al., 2009; Evgeniou & Boussios, 2005; Doumpos & Zopounidis, 2009), In the latter approach, S5-242
3 Evgeniou, Boussios and Zcharia (Evgeniou & Boussios, 2005) and Doumpos and Zopounidis (Doumpos & Zopounidis, 2009) simulated a mass of monotonic data to formulate monotonicity constraints to enforce monotonicity. One can see that the simulated data could increase the complexity of the problem computation-wise. Pelckmans et al. developed a LS- SVM regression model with monotonicity constraints. In their problems settings, instead of using simulated data, the input data are all utilized to formulate the monotonicity constraints, and they assume the input data follow a linear order and the bias term is omitted. However, such assumption may not be applied in practice. Moreover, the sparseness is lost in the LS- SVM regression model. Therefore, to deal with the shortcomings caused in the aforementioned studies, in this research, we propose a new SVR model with monotonicity constraints that are inequalities and are based on the partial order in the input data. The rest of the paper This paper is organized as follows. In Section 2, we have the related literature review. In Section 3, we discuss formulation of the monotonicity constrained SVMR model. Section 4 presents the experimental results. And finally, in Section 5, we have discussion and conclusion. LITERATURE REVIEW In this section, we review the related literatures to lay the foundation of this research project. The topics include support vector machines and classification with monotonicity constraints. Support Vector Machines SVM is the state-of-the-art neural network technology based on statistical learning (Vapnik, 1995; Vapnik, 1998). It was originally designed for binary classification in order to construct an optimal hyperplane so that the margin of separation between the negative and positive data set will be maximized. If the data are linearly separated, the optimal hyperplane will separate the data without error and the data points closest to the optimal separating hyperplane are named as support vectors. However, in practice, the data set of interest is usually linear nonseparable. In order to enhance the feasibility of linear separation, one can usually perform a non-linear transformation to the data set into a higher dimensional space, the so-called feature space. Unfortunately, the curse of dimensionality in machine learning makes the non-linear mapping too difficult to solve. SVMs solve the hurdle by using the mechanism of inner-product kernel. A comprehensive tutorial on SVM classifier has been published by Burges (1998) (Burges, 1998). Excellent performances were also obtained in the function estimation and time-series prediction applications (Müller et al., 1997; Mukherjee et al., 1997). Huang, Nakamoria, Wang (2005) (Huang et al., 2005)investigated the predictability of financial movement direction with SVM by forecasting the weekly movement direction of NIKKEI 225 index. They demonstrated that SVM outperforms Linear Discriminant Analysis, Quadratic Discriminant Analysis and Elman Backpropagation Neural Networks. Recently, SVM has S5-243
4 received much more attractions than the traditional backpropagation neural network attributed to its salient advantages (Kim & Sohn, 2010). With the advantages, many studies on SVM are presented in multiple disciplinary concerning its theory and applications. Classification with Monotonicity Constraints For classification problems with ordinal attributes very often, the class attribute should increase with each or some of the explaining attributes. They are called classification problems with the monotonicity constraints (Potharst & Feelders, 2002). The problem of classification with monotonicity constraints are commonly encountered in real-life applications such as bankruptcy risk prediction (Greco et al., 1998), finance (Gamarnik, 1998), breast cancer diagnosis (Ryu et al., 2007), house pricing (Potharst & Feelders, 2002), credit rating (Doumpos & Pasiouras, 2005) and many others. The importance of classification with monotonicity constraints had been witnessed by (Pazzani et al., 2001), which presented an evaluation work to investigate the potential for monotonicity constraints to bias machine learning systems to learn rules that were both accurate and meaningful. M. Doumpos, C. Zopounidis (Doumpos & Zopounidis, 2009) proposed a monotonic support vector machines for credit risk rating. It use the monotonicity hints to produce the virtual examples to impose the monotonic conditions which represent the special prior domain knowledge of the problem. The experimental results from a large sample of Greek industrial firms demonstrated that the introduction of the monotonicity condition reduces the danger of overfitting, thus leading to models with higher predicting ability. S. Wang applied the neural network with monotonicity property as a non-parametric efficiency analysis method to the study of efficiency analysis for private and public organizations (Wang, 2003). Simulation experiments demonstrated that their approach can remove the overhead about the parametric assumption of distribution functions as traditional efficiency analysis such as data envelopment analysis (DEA) and stochastic frontier functions (SFF). Similar work can be seen in (Pendharkar and Rodger, 2003) (Pendharkar & Rodger, 2003). Daniels and Kamp (1999) (Daniels & Kamp, 1999) proposed a monotonic neural network whose construction was done by considering multilayer neural networks with non-negative weights. In the literature we surveyed, there is a lack of method for constructing SVM with monotonicity constraints. With the increasingly more popularity ofsvm over traditional classification methods, our study isexpected to able to fulfill the need for a monotonic SVM. RESEARCH METHODOLOGY In this section, we describe the formulation of the proposed monotonicity constrained SVM model. Monotonicity is a relationship in which increasing the value of the variables always increases or decreases the likelihood category membership. Monotonicity is defined as S5-244
5 follows: Let N be the number of instances, and n be the number of attributes. Given a dataset {( ) }, with denoted as the feature space, and a partial ordering defined over this input space. A linear ordering is defined over the space Y of class values. Then the classifier is monotoneif the following statement holds: ( ) ( ) ( ) A a partial ordering on a set A is a relation which satisfies three properties containing reflexivity, anti-symmetry, and transitivity. A linear ordering is a partial order with comparability. These properties are described in the following. Reflexivity: for all Anti-symmetry: If and for any then Transitivity: If and for any then Comparability: For any either or For a set of observed dataset {( ) } the primal SVR model can be presented as ( ) ( ) ( ( ) ) Subject to { ( ( ) ) ( ) Where the inequalities ( ) ( ) for all (3) Are the monotonicity constraints. The Lagrangian for this problem is S5-245
6 ( ) ( ) ([ ( ) ] ) ( [ ( ) ] ) (4) ( ( ) ( )) ( ) with Lagrangian multipliers for and for all the optimal solution can be found at the saddle point of the Lagrangian by first minimizing over the primal variables w and b, and then maximizing over the dual multipliers and ( ) (5) Note that here by and v, we refer to ( ) and ( ). Taking the derivatives on wand b, one obtains a quadratic programming problem (dual problem) which has the following form. ( ) ( ) ( ) ( ) ( ) ( ) ( ( ) ( )) ( ) ( ( ) ( )) ( ) ( ) ( ) ( ( ) ( )) ( ) ( ) Subject to S5-246
7 ( ) A kernel trick can be applied to this quadratic form. For any symmetric, continuous function satisfying Mercer s condition [45-47], there exists a mapping ( ) such that ( ) ( ) ( ) With an appropriate choice of kernel K, the nonlinear monotonicity constrained SVM regression takes the form: Where s, and are the solution to the quadratic programming problem in (8). Now, we develop an algorithm to solve the proposed monotonicityconstrained support vector machine. Firstly, we write the objective function ( ) in matrix form. For notational convenience, we re-index and denote the new as in the following manner. Suppose there are M elements in. We can define a one-to-one mapping from the set { } to the set {( ) } and denote it as ( ). Now the optimization problem (5) can be rewritten as where G is a symmetric matrix. Apparently, the above problem is a quadraticprogramming. If the matrix G is positive S5-247
8 semidefinite, the solution is global. If G is positive definite the solution is global and unique. When G is indefinite there may exist localsolutions. Quadratic programming problems can be solved by any available quadratic programming solvers. We applied the proposed method with Gaussian kernel. The hyperparameter C and the kernel parameters can be tuned by applying k-fold cross-validation on training/validation data to a grid search. The algorithm of monotonicity constrained SVR isplanned as follows. Algorithm MC-SVR Input: Observed dataset {( ) } Output: and the corresponding classifier Steps: 1. Determine the M pairs of monotonicity constraints ( ) such that for 2. Compute the matrix G in (11). 3. Solve the quadratic programming problem in (12) for by a quadratic programming solver such as quadprog in MATLAB. 4. Apply k-fold cross-validation on training/validation data and repeat Step 3 to find the optimal parameters. 5. Output the optimal α and β Fig. 1: The MC-SVR algorithm 6. Determine the MC-SVR estimator as in (9). EXPERIMENTAL RESULTS We applied the proposed algorithm to approximate the sigmoid function on the interval [-1, 1]. In total, 100 even-spaced x values are used, along with their corresponding y values, to form the correct data set. Various perturbations, ranging from 5% to 25%, are applied to simulate artificial data. Five-fold cross-validation was adopted in the experiment. Dataset was random divided into 5 approximately equal sets, and each set was used to test the classifier, and other sets were used to train the classifier. We repeated this step until all sets have been used to test the classifier once. The Table 1 shows the comparison on accuracy of original SVR and the proposed MC-SVR. The results of artificial datasets show that the proposed monotoncity constrained SVR performs better. Table 1: Average SSE of MC-SVR and SVR on Sigmoid Data S5-248
9 CONCLUSIONS In many application areas of machine learning, there exists prior knowledge concerning the monotone relations between the response variable and predictor variables. For some cases, monotonicity is an important model requirement with a view toward explaining and justifying decisions. Therefore, in the study we propose a monotonicity constrained SVR that takes into account the monotone nature of problems. A quadratic programming problem in the dual space is developed similar to its SVR predecessor. We experimented the proposed method on the perturbed data made from Sigmoid function. The results showed that the proposed method has advantages over the original SVR. REFERENCES 1. Burges, C. J. C. (1998). Data Mining Knowledge Discovery 2, Cristianini, N. & Shawe-Taylor, J. (2000). Cambridge University Press, Ordering Information. 3. Daniels, H. & Kamp, B. (1999). Neural Computation and Applications 8, Doumpos, M. & Pasiouras, F. (2005). Computational Economics 25, Doumpos, M. & Zopounidis, C. (2009). new Mathematics and Natural Computation 5, Duivesteijn, W. & Feelders, A. (2008). Nearest Neighbour Classification with Monotonicity 7. Constraints,, 2008 European Conference on Machine Learning and Knowledge Discovery in Databases, Antwerp, Belgium Springer-Verlag. 8. Evgeniou, T., C. & Boussios, e. a. (2005). Marketing Science 24, Falck, T., Suykens, J. & De Moor, B. (2009). The 48th IEEE Conference on Decision and Control (CDC 2009) Shanghai, China 10. Gamarnik, D. (1998). In: Proceedings of the eleventh annual conference on computational learning theory, ACM Press, New York S5-249
10 11. Greco, S., Matarazzo, B. & SแowiĔski, R. (1998). In: Zopounidis, C. (ed.) Operational Tools in the Management of Financial Risks, Kluwer Academic Publishers, Dordrech Huang, W., Nakamoria, Y. & Wang, S. Y. (2005). Computers & Operations Research 32, Kim, H. S. & Sohn, S. Y. (2010). European Journal of Operational Research 201, R., Smola, A., Rไtsch, G., Sch๖lkopf, B., Kohlmorgen, J. & Vapnik, V. (1997). International Conference on Artificial Neural Networks p Springer Lecture Notes in Computer Science. 15. Mukherjee, S., Osuna, E. & Girosi, F. (1997). IEEE Workshop on Neural networks for Signal Processing 7, pp Amelia Island, FL 16. Pazzani, M. J., Mani, S. & Shankle, W. R. (2001). Methods of Information in Medicine 40, Pendharkar, P. C. & Rodger, J. A. (2003). Decision Support Systems 36, Potharst, R. & Feelders, A. J. (2002). ACM SIGKDD Explorations Newsletter 4, Ryu, Y. U., Chandrasekaran, R. & Jacob, V. (2007). European Journal of Operational Research 181, Vapnik, V. N. (1995). The Nature of Statistical Learning Theory. New York: Springer-Verlag. 21. Vapnik, V. N. (1998). Statistical Learning Theory. New York: 22. Wiley. Wang, S. (2003). Computers and Operations Research 30, 17. S5-250
Jeff Howbert Introduction to Machine Learning Winter
Classification / Regression Support Vector Machines Jeff Howbert Introduction to Machine Learning Winter 2012 1 Topics SVM classifiers for linearly separable classes SVM classifiers for non-linearly separable
More informationSupport Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM
1 Support Vector Machines (SVM) in bioinformatics Day 1: Introduction to SVM Jean-Philippe Vert Bioinformatics Center, Kyoto University, Japan Jean-Philippe.Vert@mines.org Human Genome Center, University
More informationSupport Vector Machines Explained
December 23, 2008 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
More informationSupport Vector Machine Regression for Volatile Stock Market Prediction
Support Vector Machine Regression for Volatile Stock Market Prediction Haiqin Yang, Laiwan Chan, and Irwin King Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,
More informationSupport Vector Machines
Support Vector Machines Tobias Pohlen Selected Topics in Human Language Technology and Pattern Recognition February 10, 2014 Human Language Technology and Pattern Recognition Lehrstuhl für Informatik 6
More informationSupport Vector Machine & Its Applications
Support Vector Machine & Its Applications A portion (1/3) of the slides are taken from Prof. Andrew Moore s SVM tutorial at http://www.cs.cmu.edu/~awm/tutorials Mingyue Tan The University of British Columbia
More informationSupport Vector Machines: Maximum Margin Classifiers
Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 16, 2008 Piotr Mirowski Based on slides by Sumit Chopra and Fu-Jie Huang 1 Outline What is behind
More informationCS6375: Machine Learning Gautam Kunapuli. Support Vector Machines
Gautam Kunapuli Example: Text Categorization Example: Develop a model to classify news stories into various categories based on their content. sports politics Use the bag-of-words representation for this
More informationA Tutorial on Support Vector Machine
A Tutorial on School of Computing National University of Singapore Contents Theory on Using with Other s Contents Transforming Theory on Using with Other s What is a classifier? A function that maps instances
More informationMachine Learning : Support Vector Machines
Machine Learning Support Vector Machines 05/01/2014 Machine Learning : Support Vector Machines Linear Classifiers (recap) A building block for almost all a mapping, a partitioning of the input space into
More informationVC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms
03/Feb/2010 VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms Presented by Andriy Temko Department of Electrical and Electronic Engineering Page 2 of
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction
More informationNON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION. Haiqin Yang, Irwin King and Laiwan Chan
In The Proceedings of ICONIP 2002, Singapore, 2002. NON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION Haiqin Yang, Irwin King and Laiwan Chan Department
More informationSupport'Vector'Machines. Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan
Support'Vector'Machines Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan kasthuri.kannan@nyumc.org Overview Support Vector Machines for Classification Linear Discrimination Nonlinear Discrimination
More informationBrief Introduction to Machine Learning
Brief Introduction to Machine Learning Yuh-Jye Lee Lab of Data Science and Machine Intelligence Dept. of Applied Math. at NCTU August 29, 2016 1 / 49 1 Introduction 2 Binary Classification 3 Support Vector
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Andreas Maletti Technische Universität Dresden Fakultät Informatik June 15, 2006 1 The Problem 2 The Basics 3 The Proposed Solution Learning by Machines Learning
More informationMachine Learning Support Vector Machines. Prof. Matteo Matteucci
Machine Learning Support Vector Machines Prof. Matteo Matteucci Discriminative vs. Generative Approaches 2 o Generative approach: we derived the classifier from some generative hypothesis about the way
More informationOutline. Basic concepts: SVM and kernels SVM primal/dual problems. Chih-Jen Lin (National Taiwan Univ.) 1 / 22
Outline Basic concepts: SVM and kernels SVM primal/dual problems Chih-Jen Lin (National Taiwan Univ.) 1 / 22 Outline Basic concepts: SVM and kernels Basic concepts: SVM and kernels SVM primal/dual problems
More informationNeural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science
Neural Networks Prof. Dr. Rudolf Kruse Computational Intelligence Group Faculty for Computer Science kruse@iws.cs.uni-magdeburg.de Rudolf Kruse Neural Networks 1 Supervised Learning / Support Vector Machines
More informationLinear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights
Linear Discriminant Functions and Support Vector Machines Linear, threshold units CSE19, Winter 11 Biometrics CSE 19 Lecture 11 1 X i : inputs W i : weights θ : threshold 3 4 5 1 6 7 Courtesy of University
More informationSupport Vector Machines
Support Vector Machines Stephan Dreiseitl University of Applied Sciences Upper Austria at Hagenberg Harvard-MIT Division of Health Sciences and Technology HST.951J: Medical Decision Support Overview Motivation
More informationOutliers Treatment in Support Vector Regression for Financial Time Series Prediction
Outliers Treatment in Support Vector Regression for Financial Time Series Prediction Haiqin Yang, Kaizhu Huang, Laiwan Chan, Irwin King, and Michael R. Lyu Department of Computer Science and Engineering
More informationLinear Dependency Between and the Input Noise in -Support Vector Regression
544 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 14, NO. 3, MAY 2003 Linear Dependency Between the Input Noise in -Support Vector Regression James T. Kwok Ivor W. Tsang Abstract In using the -support vector
More informationAn introduction to Support Vector Machines
1 An introduction to Support Vector Machines Giorgio Valentini DSI - Dipartimento di Scienze dell Informazione Università degli Studi di Milano e-mail: valenti@dsi.unimi.it 2 Outline Linear classifiers
More informationLecture Support Vector Machine (SVM) Classifiers
Introduction to Machine Learning Lecturer: Amir Globerson Lecture 6 Fall Semester Scribe: Yishay Mansour 6.1 Support Vector Machine (SVM) Classifiers Classification is one of the most important tasks in
More informationChapter 6 Classification and Prediction (2)
Chapter 6 Classification and Prediction (2) Outline Classification and Prediction Decision Tree Naïve Bayes Classifier Support Vector Machines (SVM) K-nearest Neighbors Accuracy and Error Measures Feature
More informationSupport Vector Machines
Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized
More informationCS798: Selected topics in Machine Learning
CS798: Selected topics in Machine Learning Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Jakramate Bootkrajang CS798: Selected topics in Machine Learning
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More informationSupport Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar
Data Mining Support Vector Machines Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar 02/03/2018 Introduction to Data Mining 1 Support Vector Machines Find a linear hyperplane
More informationLinear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction
Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the
More informationSupport Vector Machines
Support Vector Machines Some material on these is slides borrowed from Andrew Moore's excellent machine learning tutorials located at: http://www.cs.cmu.edu/~awm/tutorials/ Where Should We Draw the Line????
More informationIntroduction to SVM and RVM
Introduction to SVM and RVM Machine Learning Seminar HUS HVL UIB Yushu Li, UIB Overview Support vector machine SVM First introduced by Vapnik, et al. 1992 Several literature and wide applications Relevance
More informationNeural network time series classification of changes in nuclear power plant processes
2009 Quality and Productivity Research Conference Neural network time series classification of changes in nuclear power plant processes Karel Kupka TriloByte Statistical Research, Center for Quality and
More informationLinear Classification and SVM. Dr. Xin Zhang
Linear Classification and SVM Dr. Xin Zhang Email: eexinzhang@scut.edu.cn What is linear classification? Classification is intrinsically non-linear It puts non-identical things in the same class, so a
More informationSUPPORT VECTOR REGRESSION WITH A GENERALIZED QUADRATIC LOSS
SUPPORT VECTOR REGRESSION WITH A GENERALIZED QUADRATIC LOSS Filippo Portera and Alessandro Sperduti Dipartimento di Matematica Pura ed Applicata Universit a di Padova, Padova, Italy {portera,sperduti}@math.unipd.it
More informationA GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong
A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES Wei Chu, S. Sathiya Keerthi, Chong Jin Ong Control Division, Department of Mechanical Engineering, National University of Singapore 0 Kent Ridge Crescent,
More informationSupport Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012
Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Linear classifier Which classifier? x 2 x 1 2 Linear classifier Margin concept x 2
More informationModel Selection for LS-SVM : Application to Handwriting Recognition
Model Selection for LS-SVM : Application to Handwriting Recognition Mathias M. Adankon and Mohamed Cheriet Synchromedia Laboratory for Multimedia Communication in Telepresence, École de Technologie Supérieure,
More informationFormulation with slack variables
Formulation with slack variables Optimal margin classifier with slack variables and kernel functions described by Support Vector Machine (SVM). min (w,ξ) ½ w 2 + γσξ(i) subject to ξ(i) 0 i, d(i) (w T x(i)
More informationSupport Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature
Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature suggests the design variables should be normalized to a range of [-1,1] or [0,1].
More informationSupport Vector Machines for Classification and Regression. 1 Linearly Separable Data: Hard Margin SVMs
E0 270 Machine Learning Lecture 5 (Jan 22, 203) Support Vector Machines for Classification and Regression Lecturer: Shivani Agarwal Disclaimer: These notes are a brief summary of the topics covered in
More informationSVM TRADE-OFF BETWEEN MAXIMIZE THE MARGIN AND MINIMIZE THE VARIABLES USED FOR REGRESSION
International Journal of Pure and Applied Mathematics Volume 87 No. 6 2013, 741-750 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v87i6.2
More informationPerceptron Revisited: Linear Separators. Support Vector Machines
Support Vector Machines Perceptron Revisited: Linear Separators Binary classification can be viewed as the task of separating classes in feature space: w T x + b > 0 w T x + b = 0 w T x + b < 0 Department
More informationSUPPORT VECTOR MACHINE
SUPPORT VECTOR MACHINE Mainly based on https://nlp.stanford.edu/ir-book/pdf/15svm.pdf 1 Overview SVM is a huge topic Integration of MMDS, IIR, and Andrew Moore s slides here Our foci: Geometric intuition
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationSupport Vector Machine (continued)
Support Vector Machine continued) Overlapping class distribution: In practice the class-conditional distributions may overlap, so that the training data points are no longer linearly separable. We need
More informationReading, UK 1 2 Abstract
, pp.45-54 http://dx.doi.org/10.14257/ijseia.2013.7.5.05 A Case Study on the Application of Computational Intelligence to Identifying Relationships between Land use Characteristics and Damages caused by
More informationKernel Methods. Machine Learning A W VO
Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance
More informationNew Least Squares Support Vector Machines Based on Matrix Patterns
Neural Processing Letters (2007) 26:41 56 DOI 10.1007/s11063-007-9041-1 New Least Squares Support Vector Machines Based on Matrix Patterns Zhe Wang Songcan Chen Received: 14 August 2005 / Accepted: 7 May
More informationECE662: Pattern Recognition and Decision Making Processes: HW TWO
ECE662: Pattern Recognition and Decision Making Processes: HW TWO Purdue University Department of Electrical and Computer Engineering West Lafayette, INDIANA, USA Abstract. In this report experiments are
More informationScale-Invariance of Support Vector Machines based on the Triangular Kernel. Abstract
Scale-Invariance of Support Vector Machines based on the Triangular Kernel François Fleuret Hichem Sahbi IMEDIA Research Group INRIA Domaine de Voluceau 78150 Le Chesnay, France Abstract This paper focuses
More informationSupport Vector Machine via Nonlinear Rescaling Method
Manuscript Click here to download Manuscript: svm-nrm_3.tex Support Vector Machine via Nonlinear Rescaling Method Roman Polyak Department of SEOR and Department of Mathematical Sciences George Mason University
More informationSupport Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Support Vector Machines II CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Outline Linear SVM hard margin Linear SVM soft margin Non-linear SVM Application Linear Support Vector Machine An optimization
More informationSupport Vector Machine for Classification and Regression
Support Vector Machine for Classification and Regression Ahlame Douzal AMA-LIG, Université Joseph Fourier Master 2R - MOSIG (2013) November 25, 2013 Loss function, Separating Hyperplanes, Canonical Hyperplan
More informationSupport Vector Ordinal Regression using Privileged Information
Support Vector Ordinal Regression using Privileged Information Fengzhen Tang 1, Peter Tiňo 2, Pedro Antonio Gutiérrez 3 and Huanhuan Chen 4 1,2,4- The University of Birmingham, School of Computer Science,
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table
More informationAnnouncements - Homework
Announcements - Homework Homework 1 is graded, please collect at end of lecture Homework 2 due today Homework 3 out soon (watch email) Ques 1 midterm review HW1 score distribution 40 HW1 total score 35
More informationDiscriminative Direction for Kernel Classifiers
Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering
More informationLearning Kernel Parameters by using Class Separability Measure
Learning Kernel Parameters by using Class Separability Measure Lei Wang, Kap Luk Chan School of Electrical and Electronic Engineering Nanyang Technological University Singapore, 3979 E-mail: P 3733@ntu.edu.sg,eklchan@ntu.edu.sg
More informationNonlinear Support Vector Machines through Iterative Majorization and I-Splines
Nonlinear Support Vector Machines through Iterative Majorization and I-Splines P.J.F. Groenen G. Nalbantov J.C. Bioch July 9, 26 Econometric Institute Report EI 26-25 Abstract To minimize the primal support
More informationML (cont.): SUPPORT VECTOR MACHINES
ML (cont.): SUPPORT VECTOR MACHINES CS540 Bryan R Gibson University of Wisconsin-Madison Slides adapted from those used by Prof. Jerry Zhu, CS540-1 1 / 40 Support Vector Machines (SVMs) The No-Math Version
More information1162 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 5, SEPTEMBER The Evidence Framework Applied to Support Vector Machines
1162 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 5, SEPTEMBER 2000 Brief Papers The Evidence Framework Applied to Support Vector Machines James Tin-Yau Kwok Abstract In this paper, we show that
More informationCLUe Training An Introduction to Machine Learning in R with an example from handwritten digit recognition
CLUe Training An Introduction to Machine Learning in R with an example from handwritten digit recognition Ad Feelders Universiteit Utrecht Department of Information and Computing Sciences Algorithmic Data
More informationPattern Recognition 2018 Support Vector Machines
Pattern Recognition 2018 Support Vector Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recognition 1 / 48 Support Vector Machines Ad Feelders ( Universiteit Utrecht
More informationSupport Vector Machines.
Support Vector Machines www.cs.wisc.edu/~dpage 1 Goals for the lecture you should understand the following concepts the margin slack variables the linear support vector machine nonlinear SVMs the kernel
More informationIndirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina
Indirect Rule Learning: Support Vector Machines Indirect learning: loss optimization It doesn t estimate the prediction rule f (x) directly, since most loss functions do not have explicit optimizers. Indirection
More informationMACHINE LEARNING. Support Vector Machines. Alessandro Moschitti
MACHINE LEARNING Support Vector Machines Alessandro Moschitti Department of information and communication technology University of Trento Email: moschitti@dit.unitn.it Summary Support Vector Machines
More informationIncorporating detractors into SVM classification
Incorporating detractors into SVM classification AGH University of Science and Technology 1 2 3 4 5 (SVM) SVM - are a set of supervised learning methods used for classification and regression SVM maximal
More informationNONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition
NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function
More informationMachine Learning. Support Vector Machines. Manfred Huber
Machine Learning Support Vector Machines Manfred Huber 2015 1 Support Vector Machines Both logistic regression and linear discriminant analysis learn a linear discriminant function to separate the data
More informationPolyhedral Computation. Linear Classifiers & the SVM
Polyhedral Computation Linear Classifiers & the SVM mcuturi@i.kyoto-u.ac.jp Nov 26 2010 1 Statistical Inference Statistical: useful to study random systems... Mutations, environmental changes etc. life
More informationSupport Vector Machines
Wien, June, 2010 Paul Hofmarcher, Stefan Theussl, WU Wien Hofmarcher/Theussl SVM 1/21 Linear Separable Separating Hyperplanes Non-Linear Separable Soft-Margin Hyperplanes Hofmarcher/Theussl SVM 2/21 (SVM)
More informationStatistical Learning Reading Assignments
Statistical Learning Reading Assignments S. Gong et al. Dynamic Vision: From Images to Face Recognition, Imperial College Press, 2001 (Chapt. 3, hard copy). T. Evgeniou, M. Pontil, and T. Poggio, "Statistical
More informationLecture Notes on Support Vector Machine
Lecture Notes on Support Vector Machine Feng Li fli@sdu.edu.cn Shandong University, China 1 Hyperplane and Margin In a n-dimensional space, a hyper plane is defined by ω T x + b = 0 (1) where ω R n is
More informationSupport Vector Machines
EE 17/7AT: Optimization Models in Engineering Section 11/1 - April 014 Support Vector Machines Lecturer: Arturo Fernandez Scribe: Arturo Fernandez 1 Support Vector Machines Revisited 1.1 Strictly) Separable
More informationSupport Vector Machines for Classification and Regression
CIS 520: Machine Learning Oct 04, 207 Support Vector Machines for Classification and Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may
More informationUVA CS / Introduc8on to Machine Learning and Data Mining. Lecture 9: Classifica8on with Support Vector Machine (cont.
UVA CS 4501-001 / 6501 007 Introduc8on to Machine Learning and Data Mining Lecture 9: Classifica8on with Support Vector Machine (cont.) Yanjun Qi / Jane University of Virginia Department of Computer Science
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationSupport Vector Regression with Automatic Accuracy Control B. Scholkopf y, P. Bartlett, A. Smola y,r.williamson FEIT/RSISE, Australian National University, Canberra, Australia y GMD FIRST, Rudower Chaussee
More informationKernel Machines. Pradeep Ravikumar Co-instructor: Manuela Veloso. Machine Learning
Kernel Machines Pradeep Ravikumar Co-instructor: Manuela Veloso Machine Learning 10-701 SVM linearly separable case n training points (x 1,, x n ) d features x j is a d-dimensional vector Primal problem:
More informationSupport Vector Machines and Kernel Methods
Support Vector Machines and Kernel Methods Geoff Gordon ggordon@cs.cmu.edu July 10, 2003 Overview Why do people care about SVMs? Classification problems SVMs often produce good results over a wide range
More informationLinear classifiers selecting hyperplane maximizing separation margin between classes (large margin classifiers)
Support vector machines In a nutshell Linear classifiers selecting hyperplane maximizing separation margin between classes (large margin classifiers) Solution only depends on a small subset of training
More informationNeural networks and support vector machines
Neural netorks and support vector machines Perceptron Input x 1 Weights 1 x 2 x 3... x D 2 3 D Output: sgn( x + b) Can incorporate bias as component of the eight vector by alays including a feature ith
More informationSVMs, Duality and the Kernel Trick
SVMs, Duality and the Kernel Trick Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University February 26 th, 2007 2005-2007 Carlos Guestrin 1 SVMs reminder 2005-2007 Carlos Guestrin 2 Today
More informationMachine Learning and Data Mining. Support Vector Machines. Kalev Kask
Machine Learning and Data Mining Support Vector Machines Kalev Kask Linear classifiers Which decision boundary is better? Both have zero training error (perfect training accuracy) But, one of them seems
More informationCS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines
CS4495/6495 Introduction to Computer Vision 8C-L3 Support Vector Machines Discriminative classifiers Discriminative classifiers find a division (surface) in feature space that separates the classes Several
More informationStatistical learning theory, Support vector machines, and Bioinformatics
1 Statistical learning theory, Support vector machines, and Bioinformatics Jean-Philippe.Vert@mines.org Ecole des Mines de Paris Computational Biology group ENS Paris, november 25, 2003. 2 Overview 1.
More informationMultivariate statistical methods and data mining in particle physics Lecture 4 (19 June, 2008)
Multivariate statistical methods and data mining in particle physics Lecture 4 (19 June, 2008) RHUL Physics www.pp.rhul.ac.uk/~cowan Academic Training Lectures CERN 16 19 June, 2008 1 Outline Statement
More informationDiscussion of Some Problems About Nonlinear Time Series Prediction Using ν-support Vector Machine
Commun. Theor. Phys. (Beijing, China) 48 (2007) pp. 117 124 c International Academic Publishers Vol. 48, No. 1, July 15, 2007 Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support
More informationAn Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM
An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM Wei Chu Chong Jin Ong chuwei@gatsby.ucl.ac.uk mpeongcj@nus.edu.sg S. Sathiya Keerthi mpessk@nus.edu.sg Control Division, Department
More informationLECTURE 7 Support vector machines
LECTURE 7 Support vector machines SVMs have been used in a multitude of applications and are one of the most popular machine learning algorithms. We will derive the SVM algorithm from two perspectives:
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationBits of Machine Learning Part 1: Supervised Learning
Bits of Machine Learning Part 1: Supervised Learning Alexandre Proutiere and Vahan Petrosyan KTH (The Royal Institute of Technology) Outline of the Course 1. Supervised Learning Regression and Classification
More informationForecast daily indices of solar activity, F10.7, using support vector regression method
Research in Astron. Astrophys. 9 Vol. 9 No. 6, 694 702 http://www.raa-journal.org http://www.iop.org/journals/raa Research in Astronomy and Astrophysics Forecast daily indices of solar activity, F10.7,
More informationSupport Vector Machines. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington
Support Vector Machines CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 A Linearly Separable Problem Consider the binary classification
More informationLinear classifiers selecting hyperplane maximizing separation margin between classes (large margin classifiers)
Support vector machines In a nutshell Linear classifiers selecting hyperplane maximizing separation margin between classes (large margin classifiers) Solution only depends on a small subset of training
More informationDiscriminative Models
No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models
More informationSupport Vector Machines. Machine Learning Fall 2017
Support Vector Machines Machine Learning Fall 2017 1 Where are we? Learning algorithms Decision Trees Perceptron AdaBoost 2 Where are we? Learning algorithms Decision Trees Perceptron AdaBoost Produce
More informationUniversal Learning Technology: Support Vector Machines
Special Issue on Information Utilizing Technologies for Value Creation Universal Learning Technology: Support Vector Machines By Vladimir VAPNIK* This paper describes the Support Vector Machine (SVM) technology,
More information