Predictive Vaccinology: Optimisation of Predictions Using Support Vector Machine Classifiers

Size: px
Start display at page:

Download "Predictive Vaccinology: Optimisation of Predictions Using Support Vector Machine Classifiers"

Transcription

1 Predictive Vaccinology: Optiisation of Predictions Using Support Vector Machine Classifiers Ivana Bozic,2, Guang Lan Zhang 2,3, and Vladiir Brusic 2,4 Faculty of Matheatics, University of Belgrade, Belgrade, Serbia 2 Institute for Infoco Research, 2 Heng Mui Keng Terrace, Singapore 963 {guanglan,vladiir}@i2r.a-star.edu.sg 3 School of Coputer Engineering, Nanyang Technological University, Nanyang Avenue, Singapore School of Land and Food Sciences and the Institute for Molecular Bioscience, University of Queensland, Brisbane, Australia Abstract. Proiscuous huan leukocyte antigen (HLA) binding peptides are ideal targets for vaccine developent. Existing coputational odels for prediction of proiscuous peptides used hidden Markov odels and artificial neural networks as prediction algoriths. We report a syste based on support vector achines that outperfors previously published ethods. Preliinary testing showed that it can predict peptides binding to HLA-A2 and -A3 supertype olecules with excellent accuracy, even for olecules where no binding data are currently available. Introduction Coputational predictions of peptides (short sequences of aino acids) that bind huan leukocyte antigens (HLA) olecules of the iune syste are essential for designing vaccines and iunotherapies against cancer, infectious disease and autoiunity []. There are ore than 800 different HLA olecules characterised to date, and any of the have unique peptide binding preferences. A HLA supertype is a group of HLA olecules that share siilar olecular structure and also have siilar binding preferences they bind largely overlapping sets of peptides [2]. Soe dozen class I HLA supertypes have been described, of which four aor supertypes (HLA- A2, -A3, -B7, and -B44) are present in approxiately 90% of the huan population [2]. Prediction of peptide binding to a single HLA olecule is less relevant for the design of vaccines applicable to a large proportion of huan population. However, predictions of proiscuous peptides, i.e. those that bind ultiple HLA olecules within a supertype, are iportant in the developent of vaccines that are relevant to a broader population. Identification of peptides that bind HLA olecules is a cobinatorial proble. Therefore high accuracy of such predictions is of great iportance, because it akes identification of vaccine targets ore cost- and tie-effective. A large nuber of prediction ethods have been used for identification of HLA binding peptides. They include [reviewed in 3] binding otifs, quantitative atrices, decision trees, artificial neural networks, hidden Markov odels, and olecular odeling. In addition, it was reported that odels using support vector achines (SVM) perfor better than other prediction ethods when applied to a single HLA olecule [4,5]. Predictions of peptide binding to ultiple HLA olecules of a supertype were M. Gallagher, J. Hogan, and F. Maire (Eds.): IDEAL 2005, LNCS 3578, pp , Springer-Verlag Berlin Heidelberg 2005

2 376 Ivana Bozic, Guang Lan Zhang, and Vladiir Brusic perfored using hidden Markov odels (HMM) [6] and artificial neural networks (ANN) [7]. These reports also showed that accurate predictions proiscuous peptides within a supertype can be perfored for HLA variants for which no experiental data are available. The accuracy of predictions of peptide binding to ultiple HLA variants within A2 and A3 supertypes was easured by the area under the receiver operating characteristic curve (A ROC ) [8]. The reported values were 0.85< A ROC <0.90 for A2. We developed SVM odels for prediction of peptide binding to olecules of HLA-A2 and -A3 supertypes and copared their prediction power to the corresponding ANN and HMM odels. SVM odels use a achine learning technique that ipleents structural risk iniization principle, which iniizes the upper bound on the expected generalization error thus giving high generalization ability [9]. SVMs exhibit great resistance to over-fitting. In addition, training of SVMs is a convex prograing proble it always finds a global solution, in contrast to training of ANNs, where any local inia usually exist. 2 Data Preprocessing Nine aino acids long (9-er) peptide data were collected fro MHCPEP database [0], published articles, and a set of HLA non-binding peptides (V. Brusic unpublished data). Our datasets contain peptides with experientally verified binding affinities (binder or non-binder) for fifteen variants of HLA-A2 and eight variants of HLA- A3 supertype. Data used for training represents the interaction (contact aino acids) between a 9-er peptide and a HLA olecule. A peptide/hla interaction is represented as a virtual peptide coprising nine peptide residues plus HLA residues that coe in contact with the peptide. Only those contact residues that vary across the olecules belonging to one supertype were considered for the analysis. There are 9 non-conserved residues across A2 supertype and 4 non-conserved contact residues across A3 supertype olecules. The representation for each non-conserved residue is longer since it is a binary vector whose length equals the nuber of different residues observed at each contact position and only one position can take value. A virtual peptide for A2 supertype coprises 58 aino acids and for A3 62 aino acids. Detailed explanation of the data representation is given in [6,7,]. If A is a set of 20 aino-acids, a virtual peptide is defined as (a, a 2,,a l ), a A, {,..., l}, l = 58 for A2 and l = 62 for A3 supertype. The HLA protein sequences used in the processing of virtual peptides were extracted fro HLA Sequences Data release in the IMGT/HLA Sequence Database ( This representation enables binding peptide predictions for different olecules of the entire supertype. 3 Description of the Method 3. Support Vector Machines Let X be a training set of virtual peptides and Y be a set of their experientally deterined binding affinities. Each eber of X can have one of the two possible binding affinities ( if it is a binder, - if it is not); X = {x,, x }, x R and Y n =

3 Predictive Vaccinology: Optiisation of Predictions 377 {y,,y }, y {, }, {,..., }. We seek to find a classification function f : R n R with the property: f ( x) > 0 y = and f ( x ) < 0 y =, which would accurately predict the classes of unseen data points. The value ρ : = in f ( x ) y () is called the argin and stands for the worst classification over the whole training set. Training exaples that lie right on the argin are called support vectors. If the training data are linearly separable, we seek to find a linear function f (x) that has the axiu argin. This is equivalent to constructing a axiu-argin hyperplane that separates the two classes (binders and non-binders) in n-diensional space of our training data. In the case of linearly non-separable data, the idea is to ap the training data into a higher-diension feature space F (Hilbert space) via a non-linear ap Φ : R n F, di F > n, and construct a separating axiu-argin hyperplane there. This apping is siplified by introducing a kernel function k: k ( x, x' ) = ( Φ( x), Φ( x' )). (2) Because real-life data are usually non-separable (even in the feature space) due to noise, we allow soe isclassified training exaples. This is enabled by introduction of a new paraeter C, which is a trade-off between the argin and the training error. Constructing a axiu-argin hyperplane is a constrained convex optiization proble. Thus, solving the SVM proble is equivalent to finding a solution (see [2]) to the Karush-Kuhn-Tucker (KKT) conditions, and, equivalently, the Wolfe dual proble: Find α which axiize W(α), W ( α) = = 0 α α 2 C,, s= α α y = s y s k α y = 0. ( x, x ) This is a linearly constrained convex quadratic progra, which is solved nuerically. The decision function is = f ( x) = y α k( x, x ) + b. (4) α i are the solutions of the corresponding quadratic progra and b is easily found using the KKT copleentarity condition [2]. Coonly used kernels include linear, Gaussian and polynoial. s, (3) 3.2 Ipleentation We used the SVMlight package with a quadratic prograing tool for solving sall interediate quadratic prograing probles, based on the ethod of Hildreth and D'Espo [3]. Prior to training, every peptide fro our dataset was transfored into a

4 378 Ivana Bozic, Guang Lan Zhang, and Vladiir Brusic virtual peptide. This data were then transfored into a forat copatible with the package used, and every virtual peptide (a,a 2,,a l ) was translated into a set of n = 20 l indicators (bits). Each aino acid of a virtual peptide is represented by 20 indicators. In the coplete sequence of indicators, l indicators are set to, representing specific residues that are present at a given position, and the other 20 (l-) indicators are set to 0. In our case, x = ( i, i2,..., in ), i s {0, }, s {,..., n}, {,..., } represents a virtual peptide. The values y = ± indicate whether a peptide binds () or does not bind (-) to a HLA olecule. Blind testing was perfored for assessing the perforance of SVM for prediction of proiscuous peptides. To test the predictive accuracy of peptide binding to each of the HLA-A2 and -A3 variants, we used all peptides (binders and non-binders) related to this variant as the testing data and used all peptides related to other variants fro the sae supertype as training data. For exaple, the training set of HLA-A*020 contained all peptides related to all HLA-A2 olecules except for HLA-A*020. Testing of peptide binding to each HLA variant was perfored without inclusion of experiental data for this particular HLA variant in the training set. Testing results, therefore, are likely to represent an underestiate of the actual perforance since the final odel contains available data for all HLA variants. We perfored blind testing on five HLA-A2 and seven HLA-A3 olecules, for which sufficient data were available for valid testing (Table ). Other variants of HLA-A2 and -A3 olecules were excluded fro testing, since there was insufficient data for generating adequate test sets. Throughout blind testing, we exained three kernels (linear, Gaussian and polynoial) and various cobinations of the SVM paraeters (trade-off c, σ for Gaussian and d for polynoial kernel). We trained 50 different SVM odels for each supertype, with c varying fro 0.0 to 20, d fro to 0 and σ fro 0.00 to. Models with the best prediction perforance (highest average A ROC ) were Gaussian kernel with σ=0., c=0.5 for HLA-A2 and Gaussian kernel with σ=0. and c=2 for HLA-A3. Table. Blind testing: nuber of peptides in training and test sets for A2 and A3 supertype. Each training data set contained peptides related to other * 4 HLA-A2 or ** 7 HLA-A3 olecule variants HLA-A2 * Training data Test data HLA-A3 ** Training data Test data olecule Binders Nonbinders Binders Nonbinders olecule Binders Nonbinders Binders Nonbinders * * * * * * * * * * * * Experiental Results A ROC values of the SVM odels that showed the best prediction perforance are shown in Table 2 along with the corresponding values for optiized HMM [6] and ANN [7] predictions. The average SVM prediction accuracy is excellent for both A2 (A ROC =0.89) and A3 supertype predictions (A ROC =0.92). SVMs perfored arginally

5 Predictive Vaccinology: Optiisation of Predictions 379 better than HMM and ANN prediction ethods on A2 supertype olecules, and significantly better on A3 supertype olecules (p<0.05, Student s t-test). Table 2. A ROC values for blind testing of A2 and A3 odels; coparison with HMM and ANN HLA-A2 SVM HMM ANN HLA-A3 SVM ANN HMM olecule olecule * * * * * * * * * * * * Average Average Std.dev Std.dev Coparison of predictive perforances of all three kernels is shown in Table 3. For both A2 and A3 supertypes, the prediction odels with Gaussian kernel deostrated the best overall perforance. A ROC values for odels with polynoial kernel are very close to those of Gaussian kernels. Models using linear kernel had the lowest average A ROC on predictions for both supertypes. Table 3. Perforance coparison of SVMs with linear, polynoial and Gaussian kernels on a) HLA-A2 and b) HLA-A3. The tables show A ROC values for blind testing. The values where all three kernels perfored siilarly are shown in bold a) HLA-A2 olecule Kernel Linear Gaussian Polynoial A* A* A* A* A* Average b) HLA-A3 olecule Kernel Linear Gaussian Polynoial A* A* A* A* A* A* A* Average Discussion and Conclusion Gaussian kernel perfored best on both HLA-A2 and -A3 supertype predictions, while the linear kernel had the lowest average A ROC. This is in contrast to report [4] where the linear kernel showed the best perforance. However, the odel in [4] was

6 380 Ivana Bozic, Guang Lan Zhang, and Vladiir Brusic trained only for a single HLA olecule (A*020) and their training dataset contained saller nuber of peptides (total of 203). It is possible that linear kernel perfors well on sall datasets, because there ay not be sufficient data for fine-tuning of the boundary. Selection of an optial kernel for prediction of HLA binding peptides is dependant on the training data (both the data set and the natural properties of the studied HLA olecule). In the report [5] each of the three kernels (linear, Gaussian and polynoial) was deterined as optial for soe of the HLA olecules. In this study we report that although Gaussian kernel had the highest average A ROC, all three kernels have coparable perforances for prediction of peptide binding to olecules belonging to a HLA supertype. Possible iproveent ay be achieved by cobining predictions by SVMs that use different kernels. For exaple, iproved HLA-A3 predictions can be achieved by cobining Gaussian (e.g. A*030, 0302, 0, 02) and polynoial (e.g. A*30, 330, 680) kernels. Further experiental data are needed for validation of this hypothesis. Although linear kernel perfored best in soe cases (e.g. A*020 and 680), the iproveent relative to other kernels is iniscule. SVM showed significantly better perforance than HMM and ANN on A3 supertype olecules, but the sae conclusion could not be drawn for A2 supertype. This is probably due to the ibalance of the A2 training data with data for A*020 being significantly larger than any other A2 set. It was reported that SVM often do not have the highest accuracy on ibalanced datasets [4], and dataset for A2 supertype is ibalanced in two ways: binders/non-binders ratio is close to :4, and peptides related to A*020 constitute 6/7; peptides related to all other variants constitute only /7 of the data set. We have developed SVM odels for prediction of peptides that bind olecules belonging to HLA-A2 and -A3 supertypes. Predictions were of excellent overall accuracy for both odels. In coparison to other ethods for prediction of proiscuous peptides, HMM and ANN, SVM slightly outperfor HMM and ANN on unbalanced data sets (A2), but significantly outperfor the on balanced data sets (A3). Acknowledgeents This proect has been funded in part (ZG and VB) with the USA Federal funds fro the NIAID, NIH, Departent of Health and Huan Services, under Grant No. 5 U9 AI5654 and Contract No. HHSN C. References. Brusic,V., August,J.T.: The changing field of vaccine developent in the genoics era. Pharacogenoics 5 (2004) Sidney, J. et al: Practical, biocheical and evolutionary iplications of the discovery of HLA class I superotifs. Iunol. Today 7 (996) Brusic,V., Baic,V.B., Petrovsky,N.: Coputational ethods for prediction of T-cell epitopes-a fraework for odelling, testing, and applications. Methods 34 (2004) Zhao,Y. et al.: Application of support vector achines for T-cell epitopes prediction. Bioinforatics 9 (2003) Donnes,P. and Elofsson,A.: Prediction of MHC class I binding peptides, using SVMHC. BMC Bioinforatics 3 (2002) 25

7 Predictive Vaccinology: Optiisation of Predictions Brusic,V. et al.: Prediction of proiscuous peptides that bind HLA class I olecules, Iunol. Cell Biol. 80 (2002) Zhang,G.L. et al.: Neural odels for predicting viral vaccine targets. J. Bioinfor. Cop. Biol. (In press) 8. Swets,J.: Measuring the accuracy of diagnostic systes. Science 240 (988) Vapnik,V.N.: The Nature of Statistical Learning Theory. Springer, Berlin (995) 0. Brusic,V. et al.: MHCPEP, a database of MHC-binding peptides: update 997. Nucleic Acids Res. 26 (998) Chelvanayaga,G.: A roadap for HLA-A, HLA-B and HLA-C peptide binding specificities. Iunogenetics 45 (996) Burges,C.J.C.: A Tutorial on Support Vector Machines for Pattern Recognition. Data Mining and Knowledge Discovery 2 (998) Joachis,T.: Making Large-Scale SVM Learning Practical. Advances in Kernel Methods Support Vector Learning. MIT Press, Cabridge (999) 4. Wu,G. Chang,E.Y.: Adaptive feature-space conforal transforation for ibalanced-data learning. Proceedings of the 20 th ICML, Washington DC (2003)

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

Support Vector Machines. Maximizing the Margin

Support Vector Machines. Maximizing the Margin Support Vector Machines Support vector achines (SVMs) learn a hypothesis: h(x) = b + Σ i= y i α i k(x, x i ) (x, y ),..., (x, y ) are the training exs., y i {, } b is the bias weight. α,..., α are the

More information

Support Vector Machines. Goals for the lecture

Support Vector Machines. Goals for the lecture Support Vector Machines Mark Craven and David Page Coputer Sciences 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Soe of the slides in these lectures have been adapted/borrowed fro aterials developed

More information

Soft-margin SVM can address linearly separable problems with outliers

Soft-margin SVM can address linearly separable problems with outliers Non-linear Support Vector Machines Non-linearly separable probles Hard-argin SVM can address linearly separable probles Soft-argin SVM can address linearly separable probles with outliers Non-linearly

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

Geometrical intuition behind the dual problem

Geometrical intuition behind the dual problem Based on: Geoetrical intuition behind the dual proble KP Bennett, EJ Bredensteiner, Duality and Geoetry in SVM Classifiers, Proceedings of the International Conference on Machine Learning, 2000 1 Geoetrical

More information

Support Vector Machines. Machine Learning Series Jerry Jeychandra Blohm Lab

Support Vector Machines. Machine Learning Series Jerry Jeychandra Blohm Lab Support Vector Machines Machine Learning Series Jerry Jeychandra Bloh Lab Outline Main goal: To understand how support vector achines (SVMs) perfor optial classification for labelled data sets, also a

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,

More information

Support Vector Machines MIT Course Notes Cynthia Rudin

Support Vector Machines MIT Course Notes Cynthia Rudin Support Vector Machines MIT 5.097 Course Notes Cynthia Rudin Credit: Ng, Hastie, Tibshirani, Friedan Thanks: Şeyda Ertekin Let s start with soe intuition about argins. The argin of an exaple x i = distance

More information

UNIVERSITY OF TRENTO ON THE USE OF SVM FOR ELECTROMAGNETIC SUBSURFACE SENSING. A. Boni, M. Conci, A. Massa, and S. Piffer.

UNIVERSITY OF TRENTO ON THE USE OF SVM FOR ELECTROMAGNETIC SUBSURFACE SENSING. A. Boni, M. Conci, A. Massa, and S. Piffer. UIVRSITY OF TRTO DIPARTITO DI IGGRIA SCIZA DLL IFORAZIO 3823 Povo Trento (Italy) Via Soarive 4 http://www.disi.unitn.it O TH US OF SV FOR LCTROAGTIC SUBSURFAC SSIG A. Boni. Conci A. assa and S. Piffer

More information

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x)

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x) 7Applying Nelder Mead s Optiization Algorith APPLYING NELDER MEAD S OPTIMIZATION ALGORITHM FOR MULTIPLE GLOBAL MINIMA Abstract Ştefan ŞTEFĂNESCU * The iterative deterinistic optiization ethod could not

More information

Combining Classifiers

Combining Classifiers Cobining Classifiers Generic ethods of generating and cobining ultiple classifiers Bagging Boosting References: Duda, Hart & Stork, pg 475-480. Hastie, Tibsharini, Friedan, pg 246-256 and Chapter 10. http://www.boosting.org/

More information

Boosting with log-loss

Boosting with log-loss Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the

More information

E. Alpaydın AERFAISS

E. Alpaydın AERFAISS E. Alpaydın AERFAISS 00 Introduction Questions: Is the error rate of y classifier less than %? Is k-nn ore accurate than MLP? Does having PCA before iprove accuracy? Which kernel leads to highest accuracy

More information

PAC-Bayes Analysis Of Maximum Entropy Learning

PAC-Bayes Analysis Of Maximum Entropy Learning PAC-Bayes Analysis Of Maxiu Entropy Learning John Shawe-Taylor and David R. Hardoon Centre for Coputational Statistics and Machine Learning Departent of Coputer Science University College London, UK, WC1E

More information

A Smoothed Boosting Algorithm Using Probabilistic Output Codes

A Smoothed Boosting Algorithm Using Probabilistic Output Codes A Soothed Boosting Algorith Using Probabilistic Output Codes Rong Jin rongjin@cse.su.edu Dept. of Coputer Science and Engineering, Michigan State University, MI 48824, USA Jian Zhang jian.zhang@cs.cu.edu

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

1 Bounding the Margin

1 Bounding the Margin COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Foundations of Machine Learning Boosting. Mehryar Mohri Courant Institute and Google Research

Foundations of Machine Learning Boosting. Mehryar Mohri Courant Institute and Google Research Foundations of Machine Learning Boosting Mehryar Mohri Courant Institute and Google Research ohri@cis.nyu.edu Weak Learning Definition: concept class C is weakly PAC-learnable if there exists a (weak)

More information

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials Fast Montgoery-like Square Root Coputation over GF( ) for All Trinoials Yin Li a, Yu Zhang a, a Departent of Coputer Science and Technology, Xinyang Noral University, Henan, P.R.China Abstract This letter

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS

EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS Jochen Till, Sebastian Engell, Sebastian Panek, and Olaf Stursberg Process Control Lab (CT-AST), University of Dortund,

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis City University of New York (CUNY) CUNY Acadeic Works International Conference on Hydroinforatics 8-1-2014 Experiental Design For Model Discriination And Precise Paraeter Estiation In WDS Analysis Giovanna

More information

An improved self-adaptive harmony search algorithm for joint replenishment problems

An improved self-adaptive harmony search algorithm for joint replenishment problems An iproved self-adaptive harony search algorith for joint replenishent probles Lin Wang School of Manageent, Huazhong University of Science & Technology zhoulearner@gail.co Xiaojian Zhou School of Manageent,

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

Lecture 9: Multi Kernel SVM

Lecture 9: Multi Kernel SVM Lecture 9: Multi Kernel SVM Stéphane Canu stephane.canu@litislab.eu Sao Paulo 204 April 6, 204 Roadap Tuning the kernel: MKL The ultiple kernel proble Sparse kernel achines for regression: SVR SipleMKL:

More information

Ensemble Based on Data Envelopment Analysis

Ensemble Based on Data Envelopment Analysis Enseble Based on Data Envelopent Analysis So Young Sohn & Hong Choi Departent of Coputer Science & Industrial Systes Engineering, Yonsei University, Seoul, Korea Tel) 82-2-223-404, Fax) 82-2- 364-7807

More information

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Stochastic Subgradient Methods

Stochastic Subgradient Methods Stochastic Subgradient Methods Lingjie Weng Yutian Chen Bren School of Inforation and Coputer Science University of California, Irvine {wengl, yutianc}@ics.uci.edu Abstract Stochastic subgradient ethods

More information

A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION

A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION A eshsize boosting algorith in kernel density estiation A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION C.C. Ishiekwene, S.M. Ogbonwan and J.E. Osewenkhae Departent of Matheatics, University

More information

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair Proceedings of the 6th SEAS International Conference on Siulation, Modelling and Optiization, Lisbon, Portugal, Septeber -4, 006 0 A Siplified Analytical Approach for Efficiency Evaluation of the eaving

More information

Bayes Decision Rule and Naïve Bayes Classifier

Bayes Decision Rule and Naïve Bayes Classifier Bayes Decision Rule and Naïve Bayes Classifier Le Song Machine Learning I CSE 6740, Fall 2013 Gaussian Mixture odel A density odel p(x) ay be ulti-odal: odel it as a ixture of uni-odal distributions (e.g.

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

On Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40

On Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40 On Poset Merging Peter Chen Guoli Ding Steve Seiden Abstract We consider the follow poset erging proble: Let X and Y be two subsets of a partially ordered set S. Given coplete inforation about the ordering

More information

ANALYTICAL INVESTIGATION AND PARAMETRIC STUDY OF LATERAL IMPACT BEHAVIOR OF PRESSURIZED PIPELINES AND INFLUENCE OF INTERNAL PRESSURE

ANALYTICAL INVESTIGATION AND PARAMETRIC STUDY OF LATERAL IMPACT BEHAVIOR OF PRESSURIZED PIPELINES AND INFLUENCE OF INTERNAL PRESSURE DRAFT Proceedings of the ASME 014 International Mechanical Engineering Congress & Exposition IMECE014 Noveber 14-0, 014, Montreal, Quebec, Canada IMECE014-36371 ANALYTICAL INVESTIGATION AND PARAMETRIC

More information

MODIFICATION OF AN ANALYTICAL MODEL FOR CONTAINER LOADING PROBLEMS

MODIFICATION OF AN ANALYTICAL MODEL FOR CONTAINER LOADING PROBLEMS MODIFICATIO OF A AALYTICAL MODEL FOR COTAIER LOADIG PROBLEMS Reception date: DEC.99 otification to authors: 04 MAR. 2001 Cevriye GECER Departent of Industrial Engineering, University of Gazi 06570 Maltepe,

More information

Algorithms for parallel processor scheduling with distinct due windows and unit-time jobs

Algorithms for parallel processor scheduling with distinct due windows and unit-time jobs BULLETIN OF THE POLISH ACADEMY OF SCIENCES TECHNICAL SCIENCES Vol. 57, No. 3, 2009 Algoriths for parallel processor scheduling with distinct due windows and unit-tie obs A. JANIAK 1, W.A. JANIAK 2, and

More information

Bayesian Learning. Chapter 6: Bayesian Learning. Bayes Theorem. Roles for Bayesian Methods. CS 536: Machine Learning Littman (Wu, TA)

Bayesian Learning. Chapter 6: Bayesian Learning. Bayes Theorem. Roles for Bayesian Methods. CS 536: Machine Learning Littman (Wu, TA) Bayesian Learning Chapter 6: Bayesian Learning CS 536: Machine Learning Littan (Wu, TA) [Read Ch. 6, except 6.3] [Suggested exercises: 6.1, 6.2, 6.6] Bayes Theore MAP, ML hypotheses MAP learners Miniu

More information

Grafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space

Grafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space Journal of Machine Learning Research 3 (2003) 1333-1356 Subitted 5/02; Published 3/03 Grafting: Fast, Increental Feature Selection by Gradient Descent in Function Space Sion Perkins Space and Reote Sensing

More information

Pattern Classification using Simplified Neural Networks with Pruning Algorithm

Pattern Classification using Simplified Neural Networks with Pruning Algorithm Pattern Classification using Siplified Neural Networks with Pruning Algorith S. M. Karuzzaan 1 Ahed Ryadh Hasan 2 Abstract: In recent years, any neural network odels have been proposed for pattern classification,

More information

arxiv: v1 [cs.ds] 3 Feb 2014

arxiv: v1 [cs.ds] 3 Feb 2014 arxiv:40.043v [cs.ds] 3 Feb 04 A Bound on the Expected Optiality of Rando Feasible Solutions to Cobinatorial Optiization Probles Evan A. Sultani The Johns Hopins University APL evan@sultani.co http://www.sultani.co/

More information

Domain-Adversarial Neural Networks

Domain-Adversarial Neural Networks Doain-Adversarial Neural Networks Hana Ajakan, Pascal Gerain 2, Hugo Larochelle 3, François Laviolette 2, Mario Marchand 2,2 Départeent d inforatique et de génie logiciel, Université Laval, Québec, Canada

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3

More information

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lesson 1 4 October 2017 Outline Learning and Evaluation for Pattern Recognition Notation...2 1. The Pattern Recognition

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

The Wilson Model of Cortical Neurons Richard B. Wells

The Wilson Model of Cortical Neurons Richard B. Wells The Wilson Model of Cortical Neurons Richard B. Wells I. Refineents on the odgkin-uxley Model The years since odgkin s and uxley s pioneering work have produced a nuber of derivative odgkin-uxley-like

More information

Inspection; structural health monitoring; reliability; Bayesian analysis; updating; decision analysis; value of information

Inspection; structural health monitoring; reliability; Bayesian analysis; updating; decision analysis; value of information Cite as: Straub D. (2014). Value of inforation analysis with structural reliability ethods. Structural Safety, 49: 75-86. Value of Inforation Analysis with Structural Reliability Methods Daniel Straub

More information

Interactive Markov Models of Evolutionary Algorithms

Interactive Markov Models of Evolutionary Algorithms Cleveland State University EngagedScholarship@CSU Electrical Engineering & Coputer Science Faculty Publications Electrical Engineering & Coputer Science Departent 2015 Interactive Markov Models of Evolutionary

More information

The Simplex Method is Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate

The Simplex Method is Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate The Siplex Method is Strongly Polynoial for the Markov Decision Proble with a Fixed Discount Rate Yinyu Ye April 20, 2010 Abstract In this note we prove that the classic siplex ethod with the ost-negativereduced-cost

More information

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness A Note on Scheduling Tall/Sall Multiprocessor Tasks with Unit Processing Tie to Miniize Maxiu Tardiness Philippe Baptiste and Baruch Schieber IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights,

More information

Randomized Recovery for Boolean Compressed Sensing

Randomized Recovery for Boolean Compressed Sensing Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch

More information

Multi-view Discriminative Manifold Embedding for Pattern Classification

Multi-view Discriminative Manifold Embedding for Pattern Classification Multi-view Discriinative Manifold Ebedding for Pattern Classification X. Wang Departen of Inforation Zhenghzou 450053, China Y. Guo Departent of Digestive Zhengzhou 450053, China Z. Wang Henan University

More information

Genetic Quantum Algorithm and its Application to Combinatorial Optimization Problem

Genetic Quantum Algorithm and its Application to Combinatorial Optimization Problem Genetic Quantu Algorith and its Application to Cobinatorial Optiization Proble Kuk-Hyun Han Dept. of Electrical Engineering, KAIST, 373-, Kusong-dong Yusong-gu Taejon, 305-70, Republic of Korea khhan@vivaldi.kaist.ac.kr

More information

3D acoustic wave modeling with a time-space domain dispersion-relation-based Finite-difference scheme

3D acoustic wave modeling with a time-space domain dispersion-relation-based Finite-difference scheme P-8 3D acoustic wave odeling with a tie-space doain dispersion-relation-based Finite-difference schee Yang Liu * and rinal K. Sen State Key Laboratory of Petroleu Resource and Prospecting (China University

More information

Variational Adaptive-Newton Method

Variational Adaptive-Newton Method Variational Adaptive-Newton Method Mohaad Etiyaz Khan Wu Lin Voot Tangkaratt Zuozhu Liu Didrik Nielsen AIP, RIKEN, Tokyo Abstract We present a black-box learning ethod called the Variational Adaptive-Newton

More information

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t. CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when

More information

Lecture 21. Interior Point Methods Setup and Algorithm

Lecture 21. Interior Point Methods Setup and Algorithm Lecture 21 Interior Point Methods In 1984, Kararkar introduced a new weakly polynoial tie algorith for solving LPs [Kar84a], [Kar84b]. His algorith was theoretically faster than the ellipsoid ethod and

More information

Optimization of ripple filter for pencil beam scanning

Optimization of ripple filter for pencil beam scanning Nuclear Science and Techniques 24 (2013) 060404 Optiization of ripple filter for pencil bea scanning YANG Zhaoxia ZHANG Manzhou LI Deing * Shanghai Institute of Applied Physics, Chinese Acadey of Sciences,

More information

Optimizing energy potentials for success in protein tertiary structure prediction Ting-Lan Chiu 1 and Richard A Goldstein 1,2

Optimizing energy potentials for success in protein tertiary structure prediction Ting-Lan Chiu 1 and Richard A Goldstein 1,2 Research Paper 223 Optiizing energy potentials for success in protein tertiary structure prediction Ting-Lan Chiu 1 and Richard A Goldstein 1,2 Background: Success in solving the protein structure prediction

More information

Topic 5a Introduction to Curve Fitting & Linear Regression

Topic 5a Introduction to Curve Fitting & Linear Regression /7/08 Course Instructor Dr. Rayond C. Rup Oice: A 337 Phone: (95) 747 6958 E ail: rcrup@utep.edu opic 5a Introduction to Curve Fitting & Linear Regression EE 4386/530 Coputational ethods in EE Outline

More information

Robustness and Regularization of Support Vector Machines

Robustness and Regularization of Support Vector Machines Robustness and Regularization of Support Vector Machines Huan Xu ECE, McGill University Montreal, QC, Canada xuhuan@ci.cgill.ca Constantine Caraanis ECE, The University of Texas at Austin Austin, TX, USA

More information

Symbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm

Symbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm Acta Polytechnica Hungarica Vol., No., 04 Sybolic Analysis as Universal Tool for Deriving Properties of Non-linear Algoriths Case study of EM Algorith Vladiir Mladenović, Miroslav Lutovac, Dana Porrat

More information

Convex Programming for Scheduling Unrelated Parallel Machines

Convex Programming for Scheduling Unrelated Parallel Machines Convex Prograing for Scheduling Unrelated Parallel Machines Yossi Azar Air Epstein Abstract We consider the classical proble of scheduling parallel unrelated achines. Each job is to be processed by exactly

More information

Introduction to Discrete Optimization

Introduction to Discrete Optimization Prof. Friedrich Eisenbrand Martin Nieeier Due Date: March 9 9 Discussions: March 9 Introduction to Discrete Optiization Spring 9 s Exercise Consider a school district with I neighborhoods J schools and

More information

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words)

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words) 1 A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine (1900 words) Contact: Jerry Farlow Dept of Matheatics Univeristy of Maine Orono, ME 04469 Tel (07) 866-3540 Eail: farlow@ath.uaine.edu

More information

List Scheduling and LPT Oliver Braun (09/05/2017)

List Scheduling and LPT Oliver Braun (09/05/2017) List Scheduling and LPT Oliver Braun (09/05/207) We investigate the classical scheduling proble P ax where a set of n independent jobs has to be processed on 2 parallel and identical processors (achines)

More information

Fast Structural Similarity Search of Noncoding RNAs Based on Matched Filtering of Stem Patterns

Fast Structural Similarity Search of Noncoding RNAs Based on Matched Filtering of Stem Patterns Fast Structural Siilarity Search of Noncoding RNs Based on Matched Filtering of Ste Patterns Byung-Jun Yoon Dept. of Electrical Engineering alifornia Institute of Technology Pasadena, 91125, S Eail: bjyoon@caltech.edu

More information

HIGH RESOLUTION NEAR-FIELD MULTIPLE TARGET DETECTION AND LOCALIZATION USING SUPPORT VECTOR MACHINES

HIGH RESOLUTION NEAR-FIELD MULTIPLE TARGET DETECTION AND LOCALIZATION USING SUPPORT VECTOR MACHINES ICONIC 2007 St. Louis, O, USA June 27-29, 2007 HIGH RESOLUTION NEAR-FIELD ULTIPLE TARGET DETECTION AND LOCALIZATION USING SUPPORT VECTOR ACHINES A. Randazzo,. A. Abou-Khousa 2,.Pastorino, and R. Zoughi

More information

paper prepared for the 1996 PTRC Conference, September 2-6, Brunel University, UK ON THE CALIBRATION OF THE GRAVITY MODEL

paper prepared for the 1996 PTRC Conference, September 2-6, Brunel University, UK ON THE CALIBRATION OF THE GRAVITY MODEL paper prepared for the 1996 PTRC Conference, Septeber 2-6, Brunel University, UK ON THE CALIBRATION OF THE GRAVITY MODEL Nanne J. van der Zijpp 1 Transportation and Traffic Engineering Section Delft University

More information

The Algorithms Optimization of Artificial Neural Network Based on Particle Swarm

The Algorithms Optimization of Artificial Neural Network Based on Particle Swarm Send Orders for Reprints to reprints@benthascience.ae The Open Cybernetics & Systeics Journal, 04, 8, 59-54 59 Open Access The Algoriths Optiization of Artificial Neural Network Based on Particle Swar

More information

Recovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup)

Recovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup) Recovering Data fro Underdeterined Quadratic Measureents (CS 229a Project: Final Writeup) Mahdi Soltanolkotabi Deceber 16, 2011 1 Introduction Data that arises fro engineering applications often contains

More information

Qualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science

Qualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science Proceedings of the 6th WSEAS International Conference on Applied Coputer Science, Tenerife, Canary Islands, Spain, Deceber 16-18, 2006 183 Qualitative Modelling of Tie Series Using Self-Organizing Maps:

More information

Nonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy

Nonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy Storage Capacity and Dynaics of Nononotonic Networks Bruno Crespi a and Ignazio Lazzizzera b a. IRST, I-38050 Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I-38050 Povo (Trento) Italy INFN Gruppo

More information

Department of Electronic and Optical Engineering, Ordnance Engineering College, Shijiazhuang, , China

Department of Electronic and Optical Engineering, Ordnance Engineering College, Shijiazhuang, , China 6th International Conference on Machinery, Materials, Environent, Biotechnology and Coputer (MMEBC 06) Solving Multi-Sensor Multi-Target Assignent Proble Based on Copositive Cobat Efficiency and QPSO Algorith

More information

A method to determine relative stroke detection efficiencies from multiplicity distributions

A method to determine relative stroke detection efficiencies from multiplicity distributions A ethod to deterine relative stroke detection eiciencies ro ultiplicity distributions Schulz W. and Cuins K. 2. Austrian Lightning Detection and Inoration Syste (ALDIS), Kahlenberger Str.2A, 90 Vienna,

More information

Computational and Statistical Learning Theory

Computational and Statistical Learning Theory Coputational and Statistical Learning Theory Proble sets 5 and 6 Due: Noveber th Please send your solutions to learning-subissions@ttic.edu Notations/Definitions Recall the definition of saple based Radeacher

More information

Nyström Method vs Random Fourier Features: A Theoretical and Empirical Comparison

Nyström Method vs Random Fourier Features: A Theoretical and Empirical Comparison yströ Method vs : A Theoretical and Epirical Coparison Tianbao Yang, Yu-Feng Li, Mehrdad Mahdavi, Rong Jin, Zhi-Hua Zhou Machine Learning Lab, GE Global Research, San Raon, CA 94583 Michigan State University,

More information

Machine Learning: Fisher s Linear Discriminant. Lecture 05

Machine Learning: Fisher s Linear Discriminant. Lecture 05 Machine Learning: Fisher s Linear Discriinant Lecture 05 Razvan C. Bunescu chool of Electrical Engineering and Coputer cience bunescu@ohio.edu Lecture 05 upervised Learning ask learn an (unkon) function

More information

NBN Algorithm Introduction Computational Fundamentals. Bogdan M. Wilamoswki Auburn University. Hao Yu Auburn University

NBN Algorithm Introduction Computational Fundamentals. Bogdan M. Wilamoswki Auburn University. Hao Yu Auburn University NBN Algorith Bogdan M. Wilaoswki Auburn University Hao Yu Auburn University Nicholas Cotton Auburn University. Introduction. -. Coputational Fundaentals - Definition of Basic Concepts in Neural Network

More information

On Fuzzy Three Level Large Scale Linear Programming Problem

On Fuzzy Three Level Large Scale Linear Programming Problem J. Stat. Appl. Pro. 3, No. 3, 307-315 (2014) 307 Journal of Statistics Applications & Probability An International Journal http://dx.doi.org/10.12785/jsap/030302 On Fuzzy Three Level Large Scale Linear

More information

Predicting FTSE 100 Close Price Using Hybrid Model

Predicting FTSE 100 Close Price Using Hybrid Model SAI Intelligent Systes Conference 2015 Noveber 10-11, 2015 London, UK Predicting FTSE 100 Close Price Using Hybrid Model Bashar Al-hnaity, Departent of Electronic and Coputer Engineering, Brunel University

More information

arxiv: v1 [cs.ds] 29 Jan 2012

arxiv: v1 [cs.ds] 29 Jan 2012 A parallel approxiation algorith for ixed packing covering seidefinite progras arxiv:1201.6090v1 [cs.ds] 29 Jan 2012 Rahul Jain National U. Singapore January 28, 2012 Abstract Penghui Yao National U. Singapore

More information

Detection and Estimation Theory

Detection and Estimation Theory ESE 54 Detection and Estiation Theory Joseph A. O Sullivan Sauel C. Sachs Professor Electronic Systes and Signals Research Laboratory Electrical and Systes Engineering Washington University 11 Urbauer

More information

A Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science

A Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science A Better Algorith For an Ancient Scheduling Proble David R. Karger Steven J. Phillips Eric Torng Departent of Coputer Science Stanford University Stanford, CA 9435-4 Abstract One of the oldest and siplest

More information

Rademacher Complexity Margin Bounds for Learning with a Large Number of Classes

Rademacher Complexity Margin Bounds for Learning with a Large Number of Classes Radeacher Coplexity Margin Bounds for Learning with a Large Nuber of Classes Vitaly Kuznetsov Courant Institute of Matheatical Sciences, 25 Mercer street, New York, NY, 002 Mehryar Mohri Courant Institute

More information

SimpleMKL. hal , version 1-26 Jan Abstract. Alain Rakotomamonjy LITIS EA 4108 Université de Rouen Saint Etienne du Rouvray, France

SimpleMKL. hal , version 1-26 Jan Abstract. Alain Rakotomamonjy LITIS EA 4108 Université de Rouen Saint Etienne du Rouvray, France SipleMKL Alain Rakotoaonjy LITIS EA 48 Université de Rouen 768 Saint Etienne du Rouvray, France Francis Bach INRIA - Willow project Départeent d Inforatique, Ecole Norale Supérieure 45, Rue d Ul 7523 Paris,

More information

Protein Structure Prediction in a 210-Type Lattice Model: Parameter Optimization in the Genetic Algorithm Using Orthogonal Array

Protein Structure Prediction in a 210-Type Lattice Model: Parameter Optimization in the Genetic Algorithm Using Orthogonal Array Journal of rotein Cheistry, Vol. 8, No., 999 rotein Structure rediction in a 0-Type Lattice Model: araeter Optiization in the Genetic Algorith Using Orthogonal Array Zhirong Sun, Xiaofeng Xia, Qing Guo,

More information

A Theoretical Analysis of a Warm Start Technique

A Theoretical Analysis of a Warm Start Technique A Theoretical Analysis of a War Start Technique Martin A. Zinkevich Yahoo! Labs 701 First Avenue Sunnyvale, CA Abstract Batch gradient descent looks at every data point for every step, which is wasteful

More information

1 Proof of learning bounds

1 Proof of learning bounds COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #4 Scribe: Akshay Mittal February 13, 2013 1 Proof of learning bounds For intuition of the following theore, suppose there exists a

More information

Determining OWA Operator Weights by Mean Absolute Deviation Minimization

Determining OWA Operator Weights by Mean Absolute Deviation Minimization Deterining OWA Operator Weights by Mean Absolute Deviation Miniization Micha l Majdan 1,2 and W lodziierz Ogryczak 1 1 Institute of Control and Coputation Engineering, Warsaw University of Technology,

More information

Probability Distributions

Probability Distributions Probability Distributions In Chapter, we ephasized the central role played by probability theory in the solution of pattern recognition probles. We turn now to an exploration of soe particular exaples

More information

Estimating Parameters for a Gaussian pdf

Estimating Parameters for a Gaussian pdf Pattern Recognition and achine Learning Jaes L. Crowley ENSIAG 3 IS First Seester 00/0 Lesson 5 7 Noveber 00 Contents Estiating Paraeters for a Gaussian pdf Notation... The Pattern Recognition Proble...3

More information

W-BASED VS LATENT VARIABLES SPATIAL AUTOREGRESSIVE MODELS: EVIDENCE FROM MONTE CARLO SIMULATIONS

W-BASED VS LATENT VARIABLES SPATIAL AUTOREGRESSIVE MODELS: EVIDENCE FROM MONTE CARLO SIMULATIONS W-BASED VS LATENT VARIABLES SPATIAL AUTOREGRESSIVE MODELS: EVIDENCE FROM MONTE CARLO SIMULATIONS. Introduction When it coes to applying econoetric odels to analyze georeferenced data, researchers are well

More information