Predictive Vaccinology: Optimisation of Predictions Using Support Vector Machine Classifiers
|
|
- Diane Jenkins
- 5 years ago
- Views:
Transcription
1 Predictive Vaccinology: Optiisation of Predictions Using Support Vector Machine Classifiers Ivana Bozic,2, Guang Lan Zhang 2,3, and Vladiir Brusic 2,4 Faculty of Matheatics, University of Belgrade, Belgrade, Serbia 2 Institute for Infoco Research, 2 Heng Mui Keng Terrace, Singapore 963 {guanglan,vladiir}@i2r.a-star.edu.sg 3 School of Coputer Engineering, Nanyang Technological University, Nanyang Avenue, Singapore School of Land and Food Sciences and the Institute for Molecular Bioscience, University of Queensland, Brisbane, Australia Abstract. Proiscuous huan leukocyte antigen (HLA) binding peptides are ideal targets for vaccine developent. Existing coputational odels for prediction of proiscuous peptides used hidden Markov odels and artificial neural networks as prediction algoriths. We report a syste based on support vector achines that outperfors previously published ethods. Preliinary testing showed that it can predict peptides binding to HLA-A2 and -A3 supertype olecules with excellent accuracy, even for olecules where no binding data are currently available. Introduction Coputational predictions of peptides (short sequences of aino acids) that bind huan leukocyte antigens (HLA) olecules of the iune syste are essential for designing vaccines and iunotherapies against cancer, infectious disease and autoiunity []. There are ore than 800 different HLA olecules characterised to date, and any of the have unique peptide binding preferences. A HLA supertype is a group of HLA olecules that share siilar olecular structure and also have siilar binding preferences they bind largely overlapping sets of peptides [2]. Soe dozen class I HLA supertypes have been described, of which four aor supertypes (HLA- A2, -A3, -B7, and -B44) are present in approxiately 90% of the huan population [2]. Prediction of peptide binding to a single HLA olecule is less relevant for the design of vaccines applicable to a large proportion of huan population. However, predictions of proiscuous peptides, i.e. those that bind ultiple HLA olecules within a supertype, are iportant in the developent of vaccines that are relevant to a broader population. Identification of peptides that bind HLA olecules is a cobinatorial proble. Therefore high accuracy of such predictions is of great iportance, because it akes identification of vaccine targets ore cost- and tie-effective. A large nuber of prediction ethods have been used for identification of HLA binding peptides. They include [reviewed in 3] binding otifs, quantitative atrices, decision trees, artificial neural networks, hidden Markov odels, and olecular odeling. In addition, it was reported that odels using support vector achines (SVM) perfor better than other prediction ethods when applied to a single HLA olecule [4,5]. Predictions of peptide binding to ultiple HLA olecules of a supertype were M. Gallagher, J. Hogan, and F. Maire (Eds.): IDEAL 2005, LNCS 3578, pp , Springer-Verlag Berlin Heidelberg 2005
2 376 Ivana Bozic, Guang Lan Zhang, and Vladiir Brusic perfored using hidden Markov odels (HMM) [6] and artificial neural networks (ANN) [7]. These reports also showed that accurate predictions proiscuous peptides within a supertype can be perfored for HLA variants for which no experiental data are available. The accuracy of predictions of peptide binding to ultiple HLA variants within A2 and A3 supertypes was easured by the area under the receiver operating characteristic curve (A ROC ) [8]. The reported values were 0.85< A ROC <0.90 for A2. We developed SVM odels for prediction of peptide binding to olecules of HLA-A2 and -A3 supertypes and copared their prediction power to the corresponding ANN and HMM odels. SVM odels use a achine learning technique that ipleents structural risk iniization principle, which iniizes the upper bound on the expected generalization error thus giving high generalization ability [9]. SVMs exhibit great resistance to over-fitting. In addition, training of SVMs is a convex prograing proble it always finds a global solution, in contrast to training of ANNs, where any local inia usually exist. 2 Data Preprocessing Nine aino acids long (9-er) peptide data were collected fro MHCPEP database [0], published articles, and a set of HLA non-binding peptides (V. Brusic unpublished data). Our datasets contain peptides with experientally verified binding affinities (binder or non-binder) for fifteen variants of HLA-A2 and eight variants of HLA- A3 supertype. Data used for training represents the interaction (contact aino acids) between a 9-er peptide and a HLA olecule. A peptide/hla interaction is represented as a virtual peptide coprising nine peptide residues plus HLA residues that coe in contact with the peptide. Only those contact residues that vary across the olecules belonging to one supertype were considered for the analysis. There are 9 non-conserved residues across A2 supertype and 4 non-conserved contact residues across A3 supertype olecules. The representation for each non-conserved residue is longer since it is a binary vector whose length equals the nuber of different residues observed at each contact position and only one position can take value. A virtual peptide for A2 supertype coprises 58 aino acids and for A3 62 aino acids. Detailed explanation of the data representation is given in [6,7,]. If A is a set of 20 aino-acids, a virtual peptide is defined as (a, a 2,,a l ), a A, {,..., l}, l = 58 for A2 and l = 62 for A3 supertype. The HLA protein sequences used in the processing of virtual peptides were extracted fro HLA Sequences Data release in the IMGT/HLA Sequence Database ( This representation enables binding peptide predictions for different olecules of the entire supertype. 3 Description of the Method 3. Support Vector Machines Let X be a training set of virtual peptides and Y be a set of their experientally deterined binding affinities. Each eber of X can have one of the two possible binding affinities ( if it is a binder, - if it is not); X = {x,, x }, x R and Y n =
3 Predictive Vaccinology: Optiisation of Predictions 377 {y,,y }, y {, }, {,..., }. We seek to find a classification function f : R n R with the property: f ( x) > 0 y = and f ( x ) < 0 y =, which would accurately predict the classes of unseen data points. The value ρ : = in f ( x ) y () is called the argin and stands for the worst classification over the whole training set. Training exaples that lie right on the argin are called support vectors. If the training data are linearly separable, we seek to find a linear function f (x) that has the axiu argin. This is equivalent to constructing a axiu-argin hyperplane that separates the two classes (binders and non-binders) in n-diensional space of our training data. In the case of linearly non-separable data, the idea is to ap the training data into a higher-diension feature space F (Hilbert space) via a non-linear ap Φ : R n F, di F > n, and construct a separating axiu-argin hyperplane there. This apping is siplified by introducing a kernel function k: k ( x, x' ) = ( Φ( x), Φ( x' )). (2) Because real-life data are usually non-separable (even in the feature space) due to noise, we allow soe isclassified training exaples. This is enabled by introduction of a new paraeter C, which is a trade-off between the argin and the training error. Constructing a axiu-argin hyperplane is a constrained convex optiization proble. Thus, solving the SVM proble is equivalent to finding a solution (see [2]) to the Karush-Kuhn-Tucker (KKT) conditions, and, equivalently, the Wolfe dual proble: Find α which axiize W(α), W ( α) = = 0 α α 2 C,, s= α α y = s y s k α y = 0. ( x, x ) This is a linearly constrained convex quadratic progra, which is solved nuerically. The decision function is = f ( x) = y α k( x, x ) + b. (4) α i are the solutions of the corresponding quadratic progra and b is easily found using the KKT copleentarity condition [2]. Coonly used kernels include linear, Gaussian and polynoial. s, (3) 3.2 Ipleentation We used the SVMlight package with a quadratic prograing tool for solving sall interediate quadratic prograing probles, based on the ethod of Hildreth and D'Espo [3]. Prior to training, every peptide fro our dataset was transfored into a
4 378 Ivana Bozic, Guang Lan Zhang, and Vladiir Brusic virtual peptide. This data were then transfored into a forat copatible with the package used, and every virtual peptide (a,a 2,,a l ) was translated into a set of n = 20 l indicators (bits). Each aino acid of a virtual peptide is represented by 20 indicators. In the coplete sequence of indicators, l indicators are set to, representing specific residues that are present at a given position, and the other 20 (l-) indicators are set to 0. In our case, x = ( i, i2,..., in ), i s {0, }, s {,..., n}, {,..., } represents a virtual peptide. The values y = ± indicate whether a peptide binds () or does not bind (-) to a HLA olecule. Blind testing was perfored for assessing the perforance of SVM for prediction of proiscuous peptides. To test the predictive accuracy of peptide binding to each of the HLA-A2 and -A3 variants, we used all peptides (binders and non-binders) related to this variant as the testing data and used all peptides related to other variants fro the sae supertype as training data. For exaple, the training set of HLA-A*020 contained all peptides related to all HLA-A2 olecules except for HLA-A*020. Testing of peptide binding to each HLA variant was perfored without inclusion of experiental data for this particular HLA variant in the training set. Testing results, therefore, are likely to represent an underestiate of the actual perforance since the final odel contains available data for all HLA variants. We perfored blind testing on five HLA-A2 and seven HLA-A3 olecules, for which sufficient data were available for valid testing (Table ). Other variants of HLA-A2 and -A3 olecules were excluded fro testing, since there was insufficient data for generating adequate test sets. Throughout blind testing, we exained three kernels (linear, Gaussian and polynoial) and various cobinations of the SVM paraeters (trade-off c, σ for Gaussian and d for polynoial kernel). We trained 50 different SVM odels for each supertype, with c varying fro 0.0 to 20, d fro to 0 and σ fro 0.00 to. Models with the best prediction perforance (highest average A ROC ) were Gaussian kernel with σ=0., c=0.5 for HLA-A2 and Gaussian kernel with σ=0. and c=2 for HLA-A3. Table. Blind testing: nuber of peptides in training and test sets for A2 and A3 supertype. Each training data set contained peptides related to other * 4 HLA-A2 or ** 7 HLA-A3 olecule variants HLA-A2 * Training data Test data HLA-A3 ** Training data Test data olecule Binders Nonbinders Binders Nonbinders olecule Binders Nonbinders Binders Nonbinders * * * * * * * * * * * * Experiental Results A ROC values of the SVM odels that showed the best prediction perforance are shown in Table 2 along with the corresponding values for optiized HMM [6] and ANN [7] predictions. The average SVM prediction accuracy is excellent for both A2 (A ROC =0.89) and A3 supertype predictions (A ROC =0.92). SVMs perfored arginally
5 Predictive Vaccinology: Optiisation of Predictions 379 better than HMM and ANN prediction ethods on A2 supertype olecules, and significantly better on A3 supertype olecules (p<0.05, Student s t-test). Table 2. A ROC values for blind testing of A2 and A3 odels; coparison with HMM and ANN HLA-A2 SVM HMM ANN HLA-A3 SVM ANN HMM olecule olecule * * * * * * * * * * * * Average Average Std.dev Std.dev Coparison of predictive perforances of all three kernels is shown in Table 3. For both A2 and A3 supertypes, the prediction odels with Gaussian kernel deostrated the best overall perforance. A ROC values for odels with polynoial kernel are very close to those of Gaussian kernels. Models using linear kernel had the lowest average A ROC on predictions for both supertypes. Table 3. Perforance coparison of SVMs with linear, polynoial and Gaussian kernels on a) HLA-A2 and b) HLA-A3. The tables show A ROC values for blind testing. The values where all three kernels perfored siilarly are shown in bold a) HLA-A2 olecule Kernel Linear Gaussian Polynoial A* A* A* A* A* Average b) HLA-A3 olecule Kernel Linear Gaussian Polynoial A* A* A* A* A* A* A* Average Discussion and Conclusion Gaussian kernel perfored best on both HLA-A2 and -A3 supertype predictions, while the linear kernel had the lowest average A ROC. This is in contrast to report [4] where the linear kernel showed the best perforance. However, the odel in [4] was
6 380 Ivana Bozic, Guang Lan Zhang, and Vladiir Brusic trained only for a single HLA olecule (A*020) and their training dataset contained saller nuber of peptides (total of 203). It is possible that linear kernel perfors well on sall datasets, because there ay not be sufficient data for fine-tuning of the boundary. Selection of an optial kernel for prediction of HLA binding peptides is dependant on the training data (both the data set and the natural properties of the studied HLA olecule). In the report [5] each of the three kernels (linear, Gaussian and polynoial) was deterined as optial for soe of the HLA olecules. In this study we report that although Gaussian kernel had the highest average A ROC, all three kernels have coparable perforances for prediction of peptide binding to olecules belonging to a HLA supertype. Possible iproveent ay be achieved by cobining predictions by SVMs that use different kernels. For exaple, iproved HLA-A3 predictions can be achieved by cobining Gaussian (e.g. A*030, 0302, 0, 02) and polynoial (e.g. A*30, 330, 680) kernels. Further experiental data are needed for validation of this hypothesis. Although linear kernel perfored best in soe cases (e.g. A*020 and 680), the iproveent relative to other kernels is iniscule. SVM showed significantly better perforance than HMM and ANN on A3 supertype olecules, but the sae conclusion could not be drawn for A2 supertype. This is probably due to the ibalance of the A2 training data with data for A*020 being significantly larger than any other A2 set. It was reported that SVM often do not have the highest accuracy on ibalanced datasets [4], and dataset for A2 supertype is ibalanced in two ways: binders/non-binders ratio is close to :4, and peptides related to A*020 constitute 6/7; peptides related to all other variants constitute only /7 of the data set. We have developed SVM odels for prediction of peptides that bind olecules belonging to HLA-A2 and -A3 supertypes. Predictions were of excellent overall accuracy for both odels. In coparison to other ethods for prediction of proiscuous peptides, HMM and ANN, SVM slightly outperfor HMM and ANN on unbalanced data sets (A2), but significantly outperfor the on balanced data sets (A3). Acknowledgeents This proect has been funded in part (ZG and VB) with the USA Federal funds fro the NIAID, NIH, Departent of Health and Huan Services, under Grant No. 5 U9 AI5654 and Contract No. HHSN C. References. Brusic,V., August,J.T.: The changing field of vaccine developent in the genoics era. Pharacogenoics 5 (2004) Sidney, J. et al: Practical, biocheical and evolutionary iplications of the discovery of HLA class I superotifs. Iunol. Today 7 (996) Brusic,V., Baic,V.B., Petrovsky,N.: Coputational ethods for prediction of T-cell epitopes-a fraework for odelling, testing, and applications. Methods 34 (2004) Zhao,Y. et al.: Application of support vector achines for T-cell epitopes prediction. Bioinforatics 9 (2003) Donnes,P. and Elofsson,A.: Prediction of MHC class I binding peptides, using SVMHC. BMC Bioinforatics 3 (2002) 25
7 Predictive Vaccinology: Optiisation of Predictions Brusic,V. et al.: Prediction of proiscuous peptides that bind HLA class I olecules, Iunol. Cell Biol. 80 (2002) Zhang,G.L. et al.: Neural odels for predicting viral vaccine targets. J. Bioinfor. Cop. Biol. (In press) 8. Swets,J.: Measuring the accuracy of diagnostic systes. Science 240 (988) Vapnik,V.N.: The Nature of Statistical Learning Theory. Springer, Berlin (995) 0. Brusic,V. et al.: MHCPEP, a database of MHC-binding peptides: update 997. Nucleic Acids Res. 26 (998) Chelvanayaga,G.: A roadap for HLA-A, HLA-B and HLA-C peptide binding specificities. Iunogenetics 45 (996) Burges,C.J.C.: A Tutorial on Support Vector Machines for Pattern Recognition. Data Mining and Knowledge Discovery 2 (998) Joachis,T.: Making Large-Scale SVM Learning Practical. Advances in Kernel Methods Support Vector Learning. MIT Press, Cabridge (999) 4. Wu,G. Chang,E.Y.: Adaptive feature-space conforal transforation for ibalanced-data learning. Proceedings of the 20 th ICML, Washington DC (2003)
Kernel Methods and Support Vector Machines
Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic
More informationIntelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines
Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes
More informationSupport Vector Machines. Maximizing the Margin
Support Vector Machines Support vector achines (SVMs) learn a hypothesis: h(x) = b + Σ i= y i α i k(x, x i ) (x, y ),..., (x, y ) are the training exs., y i {, } b is the bias weight. α,..., α are the
More informationSupport Vector Machines. Goals for the lecture
Support Vector Machines Mark Craven and David Page Coputer Sciences 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Soe of the slides in these lectures have been adapted/borrowed fro aterials developed
More informationSoft-margin SVM can address linearly separable problems with outliers
Non-linear Support Vector Machines Non-linearly separable probles Hard-argin SVM can address linearly separable probles Soft-argin SVM can address linearly separable probles with outliers Non-linearly
More informationSupport Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization
Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering
More informationGeometrical intuition behind the dual problem
Based on: Geoetrical intuition behind the dual proble KP Bennett, EJ Bredensteiner, Duality and Geoetry in SVM Classifiers, Proceedings of the International Conference on Machine Learning, 2000 1 Geoetrical
More informationSupport Vector Machines. Machine Learning Series Jerry Jeychandra Blohm Lab
Support Vector Machines Machine Learning Series Jerry Jeychandra Bloh Lab Outline Main goal: To understand how support vector achines (SVMs) perfor optial classification for labelled data sets, also a
More informationFeature Extraction Techniques
Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,
More informationSupport Vector Machines MIT Course Notes Cynthia Rudin
Support Vector Machines MIT 5.097 Course Notes Cynthia Rudin Credit: Ng, Hastie, Tibshirani, Friedan Thanks: Şeyda Ertekin Let s start with soe intuition about argins. The argin of an exaple x i = distance
More informationUNIVERSITY OF TRENTO ON THE USE OF SVM FOR ELECTROMAGNETIC SUBSURFACE SENSING. A. Boni, M. Conci, A. Massa, and S. Piffer.
UIVRSITY OF TRTO DIPARTITO DI IGGRIA SCIZA DLL IFORAZIO 3823 Povo Trento (Italy) Via Soarive 4 http://www.disi.unitn.it O TH US OF SV FOR LCTROAGTIC SUBSURFAC SSIG A. Boni. Conci A. assa and S. Piffer
More informationŞtefan ŞTEFĂNESCU * is the minimum global value for the function h (x)
7Applying Nelder Mead s Optiization Algorith APPLYING NELDER MEAD S OPTIMIZATION ALGORITHM FOR MULTIPLE GLOBAL MINIMA Abstract Ştefan ŞTEFĂNESCU * The iterative deterinistic optiization ethod could not
More informationCombining Classifiers
Cobining Classifiers Generic ethods of generating and cobining ultiple classifiers Bagging Boosting References: Duda, Hart & Stork, pg 475-480. Hastie, Tibsharini, Friedan, pg 246-256 and Chapter 10. http://www.boosting.org/
More informationBoosting with log-loss
Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the
More informationE. Alpaydın AERFAISS
E. Alpaydın AERFAISS 00 Introduction Questions: Is the error rate of y classifier less than %? Is k-nn ore accurate than MLP? Does having PCA before iprove accuracy? Which kernel leads to highest accuracy
More informationPAC-Bayes Analysis Of Maximum Entropy Learning
PAC-Bayes Analysis Of Maxiu Entropy Learning John Shawe-Taylor and David R. Hardoon Centre for Coputational Statistics and Machine Learning Departent of Coputer Science University College London, UK, WC1E
More informationA Smoothed Boosting Algorithm Using Probabilistic Output Codes
A Soothed Boosting Algorith Using Probabilistic Output Codes Rong Jin rongjin@cse.su.edu Dept. of Coputer Science and Engineering, Michigan State University, MI 48824, USA Jian Zhang jian.zhang@cs.cu.edu
More informationModel Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon
Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential
More informationPattern Recognition and Machine Learning. Artificial Neural networks
Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial
More informationE0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis
E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds
More informationCOS 424: Interacting with Data. Written Exercises
COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well
More information1 Bounding the Margin
COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost
More informationPattern Recognition and Machine Learning. Artificial Neural networks
Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial
More informationFoundations of Machine Learning Boosting. Mehryar Mohri Courant Institute and Google Research
Foundations of Machine Learning Boosting Mehryar Mohri Courant Institute and Google Research ohri@cis.nyu.edu Weak Learning Definition: concept class C is weakly PAC-learnable if there exists a (weak)
More informationFast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials
Fast Montgoery-like Square Root Coputation over GF( ) for All Trinoials Yin Li a, Yu Zhang a, a Departent of Coputer Science and Technology, Xinyang Noral University, Henan, P.R.China Abstract This letter
More informatione-companion ONLY AVAILABLE IN ELECTRONIC FORM
OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer
More informationEMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS
EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS Jochen Till, Sebastian Engell, Sebastian Panek, and Olaf Stursberg Process Control Lab (CT-AST), University of Dortund,
More informationNon-Parametric Non-Line-of-Sight Identification 1
Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,
More informationExperimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis
City University of New York (CUNY) CUNY Acadeic Works International Conference on Hydroinforatics 8-1-2014 Experiental Design For Model Discriination And Precise Paraeter Estiation In WDS Analysis Giovanna
More informationAn improved self-adaptive harmony search algorithm for joint replenishment problems
An iproved self-adaptive harony search algorith for joint replenishent probles Lin Wang School of Manageent, Huazhong University of Science & Technology zhoulearner@gail.co Xiaojian Zhou School of Manageent,
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October
More informationLecture 9: Multi Kernel SVM
Lecture 9: Multi Kernel SVM Stéphane Canu stephane.canu@litislab.eu Sao Paulo 204 April 6, 204 Roadap Tuning the kernel: MKL The ultiple kernel proble Sparse kernel achines for regression: SVR SipleMKL:
More informationEnsemble Based on Data Envelopment Analysis
Enseble Based on Data Envelopent Analysis So Young Sohn & Hong Choi Departent of Coputer Science & Industrial Systes Engineering, Yonsei University, Seoul, Korea Tel) 82-2-223-404, Fax) 82-2- 364-7807
More informationIntelligent Systems: Reasoning and Recognition. Artificial Neural Networks
Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial
More informationStochastic Subgradient Methods
Stochastic Subgradient Methods Lingjie Weng Yutian Chen Bren School of Inforation and Coputer Science University of California, Irvine {wengl, yutianc}@ics.uci.edu Abstract Stochastic subgradient ethods
More informationA MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION
A eshsize boosting algorith in kernel density estiation A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION C.C. Ishiekwene, S.M. Ogbonwan and J.E. Osewenkhae Departent of Matheatics, University
More informationA Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair
Proceedings of the 6th SEAS International Conference on Siulation, Modelling and Optiization, Lisbon, Portugal, Septeber -4, 006 0 A Siplified Analytical Approach for Efficiency Evaluation of the eaving
More informationBayes Decision Rule and Naïve Bayes Classifier
Bayes Decision Rule and Naïve Bayes Classifier Le Song Machine Learning I CSE 6740, Fall 2013 Gaussian Mixture odel A density odel p(x) ay be ulti-odal: odel it as a ixture of uni-odal distributions (e.g.
More informationCS Lecture 13. More Maximum Likelihood
CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood
More informationOn Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40
On Poset Merging Peter Chen Guoli Ding Steve Seiden Abstract We consider the follow poset erging proble: Let X and Y be two subsets of a partially ordered set S. Given coplete inforation about the ordering
More informationANALYTICAL INVESTIGATION AND PARAMETRIC STUDY OF LATERAL IMPACT BEHAVIOR OF PRESSURIZED PIPELINES AND INFLUENCE OF INTERNAL PRESSURE
DRAFT Proceedings of the ASME 014 International Mechanical Engineering Congress & Exposition IMECE014 Noveber 14-0, 014, Montreal, Quebec, Canada IMECE014-36371 ANALYTICAL INVESTIGATION AND PARAMETRIC
More informationMODIFICATION OF AN ANALYTICAL MODEL FOR CONTAINER LOADING PROBLEMS
MODIFICATIO OF A AALYTICAL MODEL FOR COTAIER LOADIG PROBLEMS Reception date: DEC.99 otification to authors: 04 MAR. 2001 Cevriye GECER Departent of Industrial Engineering, University of Gazi 06570 Maltepe,
More informationAlgorithms for parallel processor scheduling with distinct due windows and unit-time jobs
BULLETIN OF THE POLISH ACADEMY OF SCIENCES TECHNICAL SCIENCES Vol. 57, No. 3, 2009 Algoriths for parallel processor scheduling with distinct due windows and unit-tie obs A. JANIAK 1, W.A. JANIAK 2, and
More informationBayesian Learning. Chapter 6: Bayesian Learning. Bayes Theorem. Roles for Bayesian Methods. CS 536: Machine Learning Littman (Wu, TA)
Bayesian Learning Chapter 6: Bayesian Learning CS 536: Machine Learning Littan (Wu, TA) [Read Ch. 6, except 6.3] [Suggested exercises: 6.1, 6.2, 6.6] Bayes Theore MAP, ML hypotheses MAP learners Miniu
More informationGrafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space
Journal of Machine Learning Research 3 (2003) 1333-1356 Subitted 5/02; Published 3/03 Grafting: Fast, Increental Feature Selection by Gradient Descent in Function Space Sion Perkins Space and Reote Sensing
More informationPattern Classification using Simplified Neural Networks with Pruning Algorithm
Pattern Classification using Siplified Neural Networks with Pruning Algorith S. M. Karuzzaan 1 Ahed Ryadh Hasan 2 Abstract: In recent years, any neural network odels have been proposed for pattern classification,
More informationarxiv: v1 [cs.ds] 3 Feb 2014
arxiv:40.043v [cs.ds] 3 Feb 04 A Bound on the Expected Optiality of Rando Feasible Solutions to Cobinatorial Optiization Probles Evan A. Sultani The Johns Hopins University APL evan@sultani.co http://www.sultani.co/
More informationDomain-Adversarial Neural Networks
Doain-Adversarial Neural Networks Hana Ajakan, Pascal Gerain 2, Hugo Larochelle 3, François Laviolette 2, Mario Marchand 2,2 Départeent d inforatique et de génie logiciel, Université Laval, Québec, Canada
More informationPattern Recognition and Machine Learning. Artificial Neural networks
Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3
More informationPattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition
Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lesson 1 4 October 2017 Outline Learning and Evaluation for Pattern Recognition Notation...2 1. The Pattern Recognition
More informationSharp Time Data Tradeoffs for Linear Inverse Problems
Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used
More informationThe Wilson Model of Cortical Neurons Richard B. Wells
The Wilson Model of Cortical Neurons Richard B. Wells I. Refineents on the odgkin-uxley Model The years since odgkin s and uxley s pioneering work have produced a nuber of derivative odgkin-uxley-like
More informationInspection; structural health monitoring; reliability; Bayesian analysis; updating; decision analysis; value of information
Cite as: Straub D. (2014). Value of inforation analysis with structural reliability ethods. Structural Safety, 49: 75-86. Value of Inforation Analysis with Structural Reliability Methods Daniel Straub
More informationInteractive Markov Models of Evolutionary Algorithms
Cleveland State University EngagedScholarship@CSU Electrical Engineering & Coputer Science Faculty Publications Electrical Engineering & Coputer Science Departent 2015 Interactive Markov Models of Evolutionary
More informationThe Simplex Method is Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate
The Siplex Method is Strongly Polynoial for the Markov Decision Proble with a Fixed Discount Rate Yinyu Ye April 20, 2010 Abstract In this note we prove that the classic siplex ethod with the ost-negativereduced-cost
More informationA Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness
A Note on Scheduling Tall/Sall Multiprocessor Tasks with Unit Processing Tie to Miniize Maxiu Tardiness Philippe Baptiste and Baruch Schieber IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights,
More informationRandomized Recovery for Boolean Compressed Sensing
Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch
More informationMulti-view Discriminative Manifold Embedding for Pattern Classification
Multi-view Discriinative Manifold Ebedding for Pattern Classification X. Wang Departen of Inforation Zhenghzou 450053, China Y. Guo Departent of Digestive Zhengzhou 450053, China Z. Wang Henan University
More informationGenetic Quantum Algorithm and its Application to Combinatorial Optimization Problem
Genetic Quantu Algorith and its Application to Cobinatorial Optiization Proble Kuk-Hyun Han Dept. of Electrical Engineering, KAIST, 373-, Kusong-dong Yusong-gu Taejon, 305-70, Republic of Korea khhan@vivaldi.kaist.ac.kr
More information3D acoustic wave modeling with a time-space domain dispersion-relation-based Finite-difference scheme
P-8 3D acoustic wave odeling with a tie-space doain dispersion-relation-based Finite-difference schee Yang Liu * and rinal K. Sen State Key Laboratory of Petroleu Resource and Prospecting (China University
More informationVariational Adaptive-Newton Method
Variational Adaptive-Newton Method Mohaad Etiyaz Khan Wu Lin Voot Tangkaratt Zuozhu Liu Didrik Nielsen AIP, RIKEN, Tokyo Abstract We present a black-box learning ethod called the Variational Adaptive-Newton
More informationThis model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.
CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when
More informationLecture 21. Interior Point Methods Setup and Algorithm
Lecture 21 Interior Point Methods In 1984, Kararkar introduced a new weakly polynoial tie algorith for solving LPs [Kar84a], [Kar84b]. His algorith was theoretically faster than the ellipsoid ethod and
More informationOptimization of ripple filter for pencil beam scanning
Nuclear Science and Techniques 24 (2013) 060404 Optiization of ripple filter for pencil bea scanning YANG Zhaoxia ZHANG Manzhou LI Deing * Shanghai Institute of Applied Physics, Chinese Acadey of Sciences,
More informationOptimizing energy potentials for success in protein tertiary structure prediction Ting-Lan Chiu 1 and Richard A Goldstein 1,2
Research Paper 223 Optiizing energy potentials for success in protein tertiary structure prediction Ting-Lan Chiu 1 and Richard A Goldstein 1,2 Background: Success in solving the protein structure prediction
More informationTopic 5a Introduction to Curve Fitting & Linear Regression
/7/08 Course Instructor Dr. Rayond C. Rup Oice: A 337 Phone: (95) 747 6958 E ail: rcrup@utep.edu opic 5a Introduction to Curve Fitting & Linear Regression EE 4386/530 Coputational ethods in EE Outline
More informationRobustness and Regularization of Support Vector Machines
Robustness and Regularization of Support Vector Machines Huan Xu ECE, McGill University Montreal, QC, Canada xuhuan@ci.cgill.ca Constantine Caraanis ECE, The University of Texas at Austin Austin, TX, USA
More informationSymbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm
Acta Polytechnica Hungarica Vol., No., 04 Sybolic Analysis as Universal Tool for Deriving Properties of Non-linear Algoriths Case study of EM Algorith Vladiir Mladenović, Miroslav Lutovac, Dana Porrat
More informationConvex Programming for Scheduling Unrelated Parallel Machines
Convex Prograing for Scheduling Unrelated Parallel Machines Yossi Azar Air Epstein Abstract We consider the classical proble of scheduling parallel unrelated achines. Each job is to be processed by exactly
More informationIntroduction to Discrete Optimization
Prof. Friedrich Eisenbrand Martin Nieeier Due Date: March 9 9 Discussions: March 9 Introduction to Discrete Optiization Spring 9 s Exercise Consider a school district with I neighborhoods J schools and
More informationA Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words)
1 A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine (1900 words) Contact: Jerry Farlow Dept of Matheatics Univeristy of Maine Orono, ME 04469 Tel (07) 866-3540 Eail: farlow@ath.uaine.edu
More informationList Scheduling and LPT Oliver Braun (09/05/2017)
List Scheduling and LPT Oliver Braun (09/05/207) We investigate the classical scheduling proble P ax where a set of n independent jobs has to be processed on 2 parallel and identical processors (achines)
More informationFast Structural Similarity Search of Noncoding RNAs Based on Matched Filtering of Stem Patterns
Fast Structural Siilarity Search of Noncoding RNs Based on Matched Filtering of Ste Patterns Byung-Jun Yoon Dept. of Electrical Engineering alifornia Institute of Technology Pasadena, 91125, S Eail: bjyoon@caltech.edu
More informationHIGH RESOLUTION NEAR-FIELD MULTIPLE TARGET DETECTION AND LOCALIZATION USING SUPPORT VECTOR MACHINES
ICONIC 2007 St. Louis, O, USA June 27-29, 2007 HIGH RESOLUTION NEAR-FIELD ULTIPLE TARGET DETECTION AND LOCALIZATION USING SUPPORT VECTOR ACHINES A. Randazzo,. A. Abou-Khousa 2,.Pastorino, and R. Zoughi
More informationpaper prepared for the 1996 PTRC Conference, September 2-6, Brunel University, UK ON THE CALIBRATION OF THE GRAVITY MODEL
paper prepared for the 1996 PTRC Conference, Septeber 2-6, Brunel University, UK ON THE CALIBRATION OF THE GRAVITY MODEL Nanne J. van der Zijpp 1 Transportation and Traffic Engineering Section Delft University
More informationThe Algorithms Optimization of Artificial Neural Network Based on Particle Swarm
Send Orders for Reprints to reprints@benthascience.ae The Open Cybernetics & Systeics Journal, 04, 8, 59-54 59 Open Access The Algoriths Optiization of Artificial Neural Network Based on Particle Swar
More informationRecovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup)
Recovering Data fro Underdeterined Quadratic Measureents (CS 229a Project: Final Writeup) Mahdi Soltanolkotabi Deceber 16, 2011 1 Introduction Data that arises fro engineering applications often contains
More informationQualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science
Proceedings of the 6th WSEAS International Conference on Applied Coputer Science, Tenerife, Canary Islands, Spain, Deceber 16-18, 2006 183 Qualitative Modelling of Tie Series Using Self-Organizing Maps:
More informationNonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy
Storage Capacity and Dynaics of Nononotonic Networks Bruno Crespi a and Ignazio Lazzizzera b a. IRST, I-38050 Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I-38050 Povo (Trento) Italy INFN Gruppo
More informationDepartment of Electronic and Optical Engineering, Ordnance Engineering College, Shijiazhuang, , China
6th International Conference on Machinery, Materials, Environent, Biotechnology and Coputer (MMEBC 06) Solving Multi-Sensor Multi-Target Assignent Proble Based on Copositive Cobat Efficiency and QPSO Algorith
More informationA method to determine relative stroke detection efficiencies from multiplicity distributions
A ethod to deterine relative stroke detection eiciencies ro ultiplicity distributions Schulz W. and Cuins K. 2. Austrian Lightning Detection and Inoration Syste (ALDIS), Kahlenberger Str.2A, 90 Vienna,
More informationComputational and Statistical Learning Theory
Coputational and Statistical Learning Theory Proble sets 5 and 6 Due: Noveber th Please send your solutions to learning-subissions@ttic.edu Notations/Definitions Recall the definition of saple based Radeacher
More informationNyström Method vs Random Fourier Features: A Theoretical and Empirical Comparison
yströ Method vs : A Theoretical and Epirical Coparison Tianbao Yang, Yu-Feng Li, Mehrdad Mahdavi, Rong Jin, Zhi-Hua Zhou Machine Learning Lab, GE Global Research, San Raon, CA 94583 Michigan State University,
More informationMachine Learning: Fisher s Linear Discriminant. Lecture 05
Machine Learning: Fisher s Linear Discriinant Lecture 05 Razvan C. Bunescu chool of Electrical Engineering and Coputer cience bunescu@ohio.edu Lecture 05 upervised Learning ask learn an (unkon) function
More informationNBN Algorithm Introduction Computational Fundamentals. Bogdan M. Wilamoswki Auburn University. Hao Yu Auburn University
NBN Algorith Bogdan M. Wilaoswki Auburn University Hao Yu Auburn University Nicholas Cotton Auburn University. Introduction. -. Coputational Fundaentals - Definition of Basic Concepts in Neural Network
More informationOn Fuzzy Three Level Large Scale Linear Programming Problem
J. Stat. Appl. Pro. 3, No. 3, 307-315 (2014) 307 Journal of Statistics Applications & Probability An International Journal http://dx.doi.org/10.12785/jsap/030302 On Fuzzy Three Level Large Scale Linear
More informationPredicting FTSE 100 Close Price Using Hybrid Model
SAI Intelligent Systes Conference 2015 Noveber 10-11, 2015 London, UK Predicting FTSE 100 Close Price Using Hybrid Model Bashar Al-hnaity, Departent of Electronic and Coputer Engineering, Brunel University
More informationarxiv: v1 [cs.ds] 29 Jan 2012
A parallel approxiation algorith for ixed packing covering seidefinite progras arxiv:1201.6090v1 [cs.ds] 29 Jan 2012 Rahul Jain National U. Singapore January 28, 2012 Abstract Penghui Yao National U. Singapore
More informationDetection and Estimation Theory
ESE 54 Detection and Estiation Theory Joseph A. O Sullivan Sauel C. Sachs Professor Electronic Systes and Signals Research Laboratory Electrical and Systes Engineering Washington University 11 Urbauer
More informationA Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science
A Better Algorith For an Ancient Scheduling Proble David R. Karger Steven J. Phillips Eric Torng Departent of Coputer Science Stanford University Stanford, CA 9435-4 Abstract One of the oldest and siplest
More informationRademacher Complexity Margin Bounds for Learning with a Large Number of Classes
Radeacher Coplexity Margin Bounds for Learning with a Large Nuber of Classes Vitaly Kuznetsov Courant Institute of Matheatical Sciences, 25 Mercer street, New York, NY, 002 Mehryar Mohri Courant Institute
More informationSimpleMKL. hal , version 1-26 Jan Abstract. Alain Rakotomamonjy LITIS EA 4108 Université de Rouen Saint Etienne du Rouvray, France
SipleMKL Alain Rakotoaonjy LITIS EA 48 Université de Rouen 768 Saint Etienne du Rouvray, France Francis Bach INRIA - Willow project Départeent d Inforatique, Ecole Norale Supérieure 45, Rue d Ul 7523 Paris,
More informationProtein Structure Prediction in a 210-Type Lattice Model: Parameter Optimization in the Genetic Algorithm Using Orthogonal Array
Journal of rotein Cheistry, Vol. 8, No., 999 rotein Structure rediction in a 0-Type Lattice Model: araeter Optiization in the Genetic Algorith Using Orthogonal Array Zhirong Sun, Xiaofeng Xia, Qing Guo,
More informationA Theoretical Analysis of a Warm Start Technique
A Theoretical Analysis of a War Start Technique Martin A. Zinkevich Yahoo! Labs 701 First Avenue Sunnyvale, CA Abstract Batch gradient descent looks at every data point for every step, which is wasteful
More information1 Proof of learning bounds
COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #4 Scribe: Akshay Mittal February 13, 2013 1 Proof of learning bounds For intuition of the following theore, suppose there exists a
More informationDetermining OWA Operator Weights by Mean Absolute Deviation Minimization
Deterining OWA Operator Weights by Mean Absolute Deviation Miniization Micha l Majdan 1,2 and W lodziierz Ogryczak 1 1 Institute of Control and Coputation Engineering, Warsaw University of Technology,
More informationProbability Distributions
Probability Distributions In Chapter, we ephasized the central role played by probability theory in the solution of pattern recognition probles. We turn now to an exploration of soe particular exaples
More informationEstimating Parameters for a Gaussian pdf
Pattern Recognition and achine Learning Jaes L. Crowley ENSIAG 3 IS First Seester 00/0 Lesson 5 7 Noveber 00 Contents Estiating Paraeters for a Gaussian pdf Notation... The Pattern Recognition Proble...3
More informationW-BASED VS LATENT VARIABLES SPATIAL AUTOREGRESSIVE MODELS: EVIDENCE FROM MONTE CARLO SIMULATIONS
W-BASED VS LATENT VARIABLES SPATIAL AUTOREGRESSIVE MODELS: EVIDENCE FROM MONTE CARLO SIMULATIONS. Introduction When it coes to applying econoetric odels to analyze georeferenced data, researchers are well
More information