Pattern Classification using Simplified Neural Networks with Pruning Algorithm

Size: px
Start display at page:

Download "Pattern Classification using Simplified Neural Networks with Pruning Algorithm"

Transcription

1 Pattern Classification using Siplified Neural Networks with Pruning Algorith S. M. Karuzzaan 1 Ahed Ryadh Hasan 2 Abstract: In recent years, any neural network odels have been proposed for pattern classification, function approxiation and regression probles. This paper presents an approach for classifying patterns fro siplified NNs. Although the predictive accuracy of ANNs is often higher than that of other ethods or huan experts, it is often said that ANNs are practically black boxes, due to the coplexity of the networks. In this paper, we have an attepted to open up these black boxes by reducing the coplexity of the network. The factor akes this possible is the pruning algorith. By eliinating redundant weights, redundant input and hidden units are identified and reoved fro the network. Using the pruning algorith, we have been able to prune networks such that only a few input units, hidden units and connections left yield a siplified network. Experiental results on several bencharks probles in neural networks show the effectiveness of the proposed approach with good generalization ability. Keywords: Artificial Neural Network, Pattern Classification, Pruning Algorith, Weight Eliination, Penalty Function, Network Siplification. 1 Introduction In recent years, any neural network odels have been proposed for pattern classification, function approxiation and regression probles [2] [3] [18]. Aong the, the class of ulti-layer feed forward networks is ost popular. Methods using standard back propagation perfor gradient descent only in the weight space of a network with fixed topology [13]. In general, this approach is useful only when the network architecture is chosen correctly [9]. Too sall a network cannot learn the proble well or too large a size will lead to over fitting and poor generalization [1]. Artificial neural networks are considered as efficient coputing odels and as the universal approxiators [4]. The predictive accuracy of neural network is higher than that of other ethods or huan experts, it is generally difficult to understand how the network arrives at a particular decision due to the coplexity of a particular architecture [6] [15]. One of the ajor criticis is their being black boxes, since no satisfactory explanation of their behavior has been offered. This is because of the coplexity of the interconnections between layers and the network size [18]. As such, an optial network size with inial nuber of interconnection will give insight into how neural network perfors. Another otivation for network siplification and pruning is related to tie coplexity of learning tie [7] [8]. 2 Pruning Algorith Network pruning offers another approach for dynaically deterining an appropriate network topology. Pruning techniques [11] begin by training a larger than necessary network and then eliinate weights and neurons that are deeed redundant. Typically, ethods for reoving weights involve adding a penalty ter to the error function [5]. It is hoped that adding a penalty ter to the error function, unnecessary connection will have saller weights and therefore coplexity of the network can be significantly reduced. This paper ais at pruning the 1 Assistant Professor, Departent of Coputer Science and Engineering, Manarat International University, Dhaka- 1212, Bangladesh, Eail: sk_iiuc@yahoo.co 2 School of Counication, Independent University Bangladesh, Chittagong, Bangladesh. Eail: ryadh78@yahoo.co

2 network size both in nuber of neurons and nuber of interconnections between the neurons. The pruning strategies along with the penalty function are described in the subsequent sections. 2.1 Penalty Function When a network is to be pruned, it is a coon practice to add a penalty ter to the error function during training [16]. Usually, the penalty ter, as suggested in different literature, is 2 2 h n β ( w h o h n h o l ) β ( v ) p 2 2 P( w, v) = ε ε ( wl ) + ( vp ) 1 l 11 β ( w ) 1 p 11 ( ) = = 1 l 1 1 p 1 l = = β v (1) + + p = = = = Given an n-diensional exaple i x, iε { 1, 2,..., k} as input, let wl be the weight for the connection fro input unit ll, ε { 1, 2,..., n} to hidden unit, ε { 1, 2,..., h} unit to output unit p, pε { 1, 2,..., o} and v p be the weight for the connection fro hidden, the p th output of the network for exaple x i is obtained by coputing h i Sp = σ α vp, where = 1 (2) n i α = δ xlwl, ( ) ( x )/ ( δ x = e e e e x ) l= 1 (3) i The target output fro an exaple x that belongs to class C j is an o-diensional vector t i, where t i p = 0 if p = j i and t p = 1, j, p = 1,2 o. The back propagation algorith is applied to update the weights (w, v) and iniize the following function: θ wv, = F wv, + P wv, (4) ( ) ( ) ( ) where F( w, v ) is the cross entropy function as defined k o i i i i (, ) = ( plog p + (1 p)log(1 p) ) F w v t S t S (5) i= 1 p= 1 and P( w, v ) is a penalty ter as described in (1) used for weight decay. 2.2 Redundant weight Pruning Penalty function is used for weight decay. As such we can eliinate redundant weights with the following Weight Eliination Algorith as suggested in different literature [12][14][17] Weight Eliination Algorith: 1. Let η 1 and η 2 be positive scalars such that η 1 + η 2 < Pick a fully connected network and train this network such that error condition is satisfied by all input patterns. Let (w, v) be the weights of this network. w l 3. For each, if v w l 4 η 2 (6) Then reove wl fro the network

3 4. For each v, If v 4 η 2 (7) Then reove v fro the network 5. If no weight satisfies condition (6) or condition (7) then reove v w l. wl with the sallest product 6. Retrain the network. If classification rate of the network falls below an acceptable level, then stop. Otherwise go to Step Input and Hidden Node Pruning A node-pruning algorith is presented below to reove redundant nodes in the input and hidden layer Input and Hidden node Pruning Algorith: Step 1: Create an initial network with as any input neurons as required by the specific proble description and with one hidden unit. Randoly initialize the connection weights of the network within a certain range. Step 2: Partially train the network on the training set for a certain nuber of training epochs using a training algorith. The nuber of training epochs, τ, is specified by the user. Step 3: Eliinate the redundant weights by using weight eliination algorith as described in section 2.2. Step 4: Test this network. If the accuracy of this network falls below an acceptable range then add one ore hidden unit and go to step 2. Step 5: If there is any input node x l with w l = 0, for = 1,2 h, then reove this node. Step 6: Test the generalization ability of the network with test set. If the network successfully converges then erinate, otherwise, go to step Experiental Results And Discussions In this experient, we have used three benchark classification probles. The probles are breast cancer diagnosis, classification of glass types and Pia Indians Diabetes diagnosis proble [10] [19]. All the data sets were obtained fro the UCI achine learning benchark repository. Brief characteristics of the data sets are listed in Table 1. Table 1: Characteristics of data sets. Data set Input Output Output Training Validation Test Total Attributes Units Classes Exaples Exaples exaples exaples Cancer Glass Diabetes The experiental results of different data sets are shown in table 2, figure 1, 2 and 3. In the experiental results of cancer data set, we have found that a fully connected network of architecture has the classification accuracy of

4 97.143%. After pruning the network with Weight Eliination Algorith and Input and Hidden node Pruning Algorith, we have found a siplified network of architecture with classification accuracy of %. The graphical representation of the siplified network is given in figure 3. It shows that only the input attributes I 1, I 6, I 9 along with a single hidden unit is adequate for this proble. Output Layer Hidden Layer O 1 O 2 W 1 = W 6 = W 9 = V 1 = V 2 = Input Layer I 1 I 2 I 3 I 4 I 5 I 6 I 7 I 8 I 9 W i = Input to Hidden Weight V i = Hidden to Output Weight I i = Input signal O i = Output Signal Active Weight Pruned Weight Active Neuron Pruned Neuron Figure 1: Siplified Network for Breast Cancer Diagnosis proble. O 1 O 2 W12 = W 13 = Output Layer W 14 = W 18 = W 21 = W 23 = Hidden Layer W 24 = W 25 = W 26 = W 27 = Input Layer W 28 = I 1 I 2 I 3 I 4 I 5 I 6 I 7 I 8 V 11 = V 12 = V 21 = V 22 = W i = Input to Hidden Weight V i = Hidden to Output Weight Active Weight Pruned Weight I i = Input signal Active Neuron O i = Output Signal Pruned Neuron Figure 2: Siplified Network for Pia Indians Diabetes diagnosis proble.

5 In the experiental results of Pia Indians Diabetes data set, we have found that a fully connected network of architecture has the classification accuracy of %. After pruning the network with Weight Eliination Algorith and Input and Hidden node Pruning Algorith, we have found a siplified network of architecture with classification accuracy of %. The graphical representation of the siplified network is given in figure 2. It shows that no input attribute can be reoved but a hidden node along with soe redundant connection has been reoved which have been shown with a dotted line in figure 2. O 1 O 2 O 3 O 4 O 5 O 6 Output Layer Hidden Layer Input Layer I 1 I 2 I 3 I 4 I 5 I 6 I 7 I 8 I 9 I i = Input signal O i = Output Signal Active Weight Pruned Weight Active Neuron Pruned Neuron Figure 3: Siplified Network for Glass classification proble. In the experiental results of Glass classification data set, we have found that a fully connected network of architecture has the classification accuracy of %. After pruning the network with Weight Eliination Algorith and Input and Hidden node Pruning Algorith, we have found a siplified network of architecture with classification accuracy of %. The graphical representation of the siplified network is given in figure 3. It shows that no input attribute can be reoved but a hidden node along with soe redundant connection has been reoved which have been shown with a dotted line in figure 3.

6 Table 2: Experiental Results Data sets Cancer1 Diabetes Glass Results Learning Rate No. of Epoch Initial Architecture Input Nodes Reoved Hidden Nodes Reoved Total Connection Reoved Siplified Architecture Accuracy (%) of fully connected network Accuracy (%) of siplified network Fro the experiental results discussed above, it can be said that not all input attributes and weights are equally iportant. Moreover, it is difficult to deterine the appropriate nuber of hidden nodes. By pruning approach we can autoatically deterine an appropriate nuber of hidden nodes. We can reove redundant nodes and connections without sacrificing significant accuracy using network pruning approach discussed in section 2.2 and section 2.3. As such we can reduce coputational cost by using the siplified networks. 4 Future Work In future we will use this network pruning approach for rule extraction and feature selection. These pruning strategies will be also exained for function approxiation and regression probles. 5 Conclusions In this paper we proposed an efficient network siplification algorith using pruning strategies. Using this approach we obtain optial network architecture with inial nuber of connections and neurons without deteriorating the perforance of the network significantly. Experiental results show that the perforance of the siplified network is quite significant and acceptable copared to fully connected network. This siplification of the network ensures both reliability and reduced coputational cost. Reference [1] T. Ash, Dynaic node creation in backpropagation networks, Connection Sci., vol. 1, pp , [2] R. W. Brause, Medical Analysis and Diagnosis by Neural Networks, J.W. Goethe-University, Coputer Science Dept., Frankfurt a. M., Gerany. [3] J. W., Everhart, J. E., Dickson, W. C., Knowler, W. C., Johannes, R. S., Using the ADAP learning algorith to forecast the onset of diabetes ellitus, Proc. Syp. on Coputer Applications and Medical Care (Piscataway, NJ: IEEE Coputer Society Press), pp , 1988.

7 [4] S. E. Fahlan and C. Lebiere, The cascade-correlation learning architecture, in Advances in Neural Inforation Processing Syste 2, D. S. Touretzky, Ed. San Mateo, CA: Morgan Kaufann, pp , [5] Sion Haykin, Neural Networks- A Coprehensive Foundation, Second Edition, Pearson Edition Asia, Third Indian Reprint, [6] T. Y. Kwok and D. Y. Yeung, Constructive Algorith for Structure Learning in feed- forward neural network for regression probles, IEEE Trans. Neural Networks, vol. 8, pp , [7] M. Monirul. Isla and K. Murase, A new algorith to design copact two hidden-layer artificial neural networks, Neural Networks, vol. 4, pp , [8] M. Monirul Isla, M. A. H. Akhand, M. Abdur Rahan and K. Murase, Weight Freezing to Reduce Training Tie in Designing Artificial neural Networks, Proceedings of 5 th ICCIT, EWU, pp , Deceber [9] R. Parekh, J.Yang, and V. Honavar, Constructive Neural Network Learning Algoriths for Pattern Classification, IEEE Trans. Neural Networks, vol. 11, no. 2, March [10] L. Prechelt, Proben1-A Set of Neural Network Benchark Probles and Bencharking Rules, University of Karlsruhe, Gerany, [11] R. Reed, Pruning algoriths-a survey, IEEE Trans. Neural Networks, vol. 4, pp , [12] R. Setiono and L.C.K. Hui, Use of quasi-newton ethod in a feedforward neural network construction algorith, IEEE Trans. Neural Networks, vol. 6, no.1, pp , Jan [13] R. Setiono, Huan Liu, Understanding Neural networks via Rule Extraction, In Proceedings of the International Joint conference on Artificial Intelligence, pp , [14] R. Setiono, Huan Liu, Iproving Backpropagation Learning with Feature Selection, Applied Intelligence, vol. 6, no. 2, pp , [15] R. Setiono, Extracting rules fro pruned networks for breast cancer diagnosis, Artificial Intelligence in Machine, vol. 8, no. 1, pp , [16] R. Setiono, A penalty function approach for pruning Feedforward neural networks, Neural Coputation, vol. 9, no. 1, pp , [17] R. Setiono, Techniques for extracting rules fro artificial neural networks, Plenary Lecture presented at the 5th International Conference on Soft Coputing and Inforation Systes, Iizuka, Japan, October [18] R. Setiono W. K. Leow and J. M. Zurada, Extraction of rules fro artificial neural networks for nonlinear regression, IEEE Trans. Neural Networks, vol. 13, no.3, pp , [19] W. H. Wolberg and O.L. Mangasarian, Multisurface ethod of pattern separation for edical diagnosis applied to breast cytology, Proceedings of the National Acadey of Sciences, U SA, Volue 87, pp , Deceber 1990.

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

CHARACTER RECOGNITION USING A SELF-ADAPTIVE TRAINING

CHARACTER RECOGNITION USING A SELF-ADAPTIVE TRAINING CHARACTER RECOGNITION USING A SELF-ADAPTIVE TRAINING Dr. Eng. Shasuddin Ahed $ College of Business and Econoics (AACSB Accredited) United Arab Eirates University, P O Box 7555, Al Ain, UAE. and $ Edith

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Bayes Decision Rule and Naïve Bayes Classifier

Bayes Decision Rule and Naïve Bayes Classifier Bayes Decision Rule and Naïve Bayes Classifier Le Song Machine Learning I CSE 6740, Fall 2013 Gaussian Mixture odel A density odel p(x) ay be ulti-odal: odel it as a ixture of uni-odal distributions (e.g.

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial

More information

NBN Algorithm Introduction Computational Fundamentals. Bogdan M. Wilamoswki Auburn University. Hao Yu Auburn University

NBN Algorithm Introduction Computational Fundamentals. Bogdan M. Wilamoswki Auburn University. Hao Yu Auburn University NBN Algorith Bogdan M. Wilaoswki Auburn University Hao Yu Auburn University Nicholas Cotton Auburn University. Introduction. -. Coputational Fundaentals - Definition of Basic Concepts in Neural Network

More information

The Algorithms Optimization of Artificial Neural Network Based on Particle Swarm

The Algorithms Optimization of Artificial Neural Network Based on Particle Swarm Send Orders for Reprints to reprints@benthascience.ae The Open Cybernetics & Systeics Journal, 04, 8, 59-54 59 Open Access The Algoriths Optiization of Artificial Neural Network Based on Particle Swar

More information

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis City University of New York (CUNY) CUNY Acadeic Works International Conference on Hydroinforatics 8-1-2014 Experiental Design For Model Discriination And Precise Paraeter Estiation In WDS Analysis Giovanna

More information

Ensemble Based on Data Envelopment Analysis

Ensemble Based on Data Envelopment Analysis Enseble Based on Data Envelopent Analysis So Young Sohn & Hong Choi Departent of Coputer Science & Industrial Systes Engineering, Yonsei University, Seoul, Korea Tel) 82-2-223-404, Fax) 82-2- 364-7807

More information

Multilayer Neural Networks

Multilayer Neural Networks Multilayer Neural Networs Brain s. Coputer Designed to sole logic and arithetic probles Can sole a gazillion arithetic and logic probles in an hour absolute precision Usually one ery fast procesor high

More information

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness A Note on Scheduling Tall/Sall Multiprocessor Tasks with Unit Processing Tie to Miniize Maxiu Tardiness Philippe Baptiste and Baruch Schieber IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights,

More information

Combining Classifiers

Combining Classifiers Cobining Classifiers Generic ethods of generating and cobining ultiple classifiers Bagging Boosting References: Duda, Hart & Stork, pg 475-480. Hastie, Tibsharini, Friedan, pg 246-256 and Chapter 10. http://www.boosting.org/

More information

Hand Written Digit Recognition Using Backpropagation Neural Network on Master-Slave Architecture

Hand Written Digit Recognition Using Backpropagation Neural Network on Master-Slave Architecture J.V.S.Srinivas et al./ International Journal of Coputer Science & Engineering Technology (IJCSET) Hand Written Digit Recognition Using Backpropagation Neural Network on Master-Slave Architecture J.V.S.SRINIVAS

More information

IAENG International Journal of Computer Science, 42:2, IJCS_42_2_06. Approximation Capabilities of Interpretable Fuzzy Inference Systems

IAENG International Journal of Computer Science, 42:2, IJCS_42_2_06. Approximation Capabilities of Interpretable Fuzzy Inference Systems IAENG International Journal of Coputer Science, 4:, IJCS_4 6 Approxiation Capabilities of Interpretable Fuzzy Inference Systes Hirofui Miyajia, Noritaka Shigei, and Hiroi Miyajia 3 Abstract Many studies

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

ZISC Neural Network Base Indicator for Classification Complexity Estimation

ZISC Neural Network Base Indicator for Classification Complexity Estimation ZISC Neural Network Base Indicator for Classification Coplexity Estiation Ivan Budnyk, Abdennasser Сhebira and Kurosh Madani Iages, Signals and Intelligent Systes Laboratory (LISSI / EA 3956) PARIS XII

More information

A Novel Breast Mass Diagnosis System based on Zernike Moments as Shape and Density Descriptors

A Novel Breast Mass Diagnosis System based on Zernike Moments as Shape and Density Descriptors A Novel Breast Mass Diagnosis Syste based on Zernike Moents as Shape and Density Descriptors Air Tahasbi 1,2, *, Fateeh Saki 2, Haed Aghapanah 3, Shahriar B. Shokouhi 2 1 Departent of Electrical Engineering,

More information

1 Proof of learning bounds

1 Proof of learning bounds COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #4 Scribe: Akshay Mittal February 13, 2013 1 Proof of learning bounds For intuition of the following theore, suppose there exists a

More information

Support Vector Machines. Goals for the lecture

Support Vector Machines. Goals for the lecture Support Vector Machines Mark Craven and David Page Coputer Sciences 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Soe of the slides in these lectures have been adapted/borrowed fro aterials developed

More information

An improved self-adaptive harmony search algorithm for joint replenishment problems

An improved self-adaptive harmony search algorithm for joint replenishment problems An iproved self-adaptive harony search algorith for joint replenishent probles Lin Wang School of Manageent, Huazhong University of Science & Technology zhoulearner@gail.co Xiaojian Zhou School of Manageent,

More information

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x)

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x) 7Applying Nelder Mead s Optiization Algorith APPLYING NELDER MEAD S OPTIMIZATION ALGORITHM FOR MULTIPLE GLOBAL MINIMA Abstract Ştefan ŞTEFĂNESCU * The iterative deterinistic optiization ethod could not

More information

Qualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science

Qualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science Proceedings of the 6th WSEAS International Conference on Applied Coputer Science, Tenerife, Canary Islands, Spain, Deceber 16-18, 2006 183 Qualitative Modelling of Tie Series Using Self-Organizing Maps:

More information

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann ECLT 5810 Classification Neural Networks Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann Neural Networks A neural network is a set of connected input/output

More information

PAC-Bayes Analysis Of Maximum Entropy Learning

PAC-Bayes Analysis Of Maximum Entropy Learning PAC-Bayes Analysis Of Maxiu Entropy Learning John Shawe-Taylor and David R. Hardoon Centre for Coputational Statistics and Machine Learning Departent of Coputer Science University College London, UK, WC1E

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

Soft-margin SVM can address linearly separable problems with outliers

Soft-margin SVM can address linearly separable problems with outliers Non-linear Support Vector Machines Non-linearly separable probles Hard-argin SVM can address linearly separable probles Soft-argin SVM can address linearly separable probles with outliers Non-linearly

More information

Block designs and statistics

Block designs and statistics Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3

More information

A Smoothed Boosting Algorithm Using Probabilistic Output Codes

A Smoothed Boosting Algorithm Using Probabilistic Output Codes A Soothed Boosting Algorith Using Probabilistic Output Codes Rong Jin rongjin@cse.su.edu Dept. of Coputer Science and Engineering, Michigan State University, MI 48824, USA Jian Zhang jian.zhang@cs.cu.edu

More information

Support Vector Machines. Maximizing the Margin

Support Vector Machines. Maximizing the Margin Support Vector Machines Support vector achines (SVMs) learn a hypothesis: h(x) = b + Σ i= y i α i k(x, x i ) (x, y ),..., (x, y ) are the training exs., y i {, } b is the bias weight. α,..., α are the

More information

Predicting FTSE 100 Close Price Using Hybrid Model

Predicting FTSE 100 Close Price Using Hybrid Model SAI Intelligent Systes Conference 2015 Noveber 10-11, 2015 London, UK Predicting FTSE 100 Close Price Using Hybrid Model Bashar Al-hnaity, Departent of Electronic and Coputer Engineering, Brunel University

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

GREY FORECASTING AND NEURAL NETWORK MODEL OF SPORT PERFORMANCE

GREY FORECASTING AND NEURAL NETWORK MODEL OF SPORT PERFORMANCE Journal of heoretical and Applied Inforation echnology 3 st March 03 Vol 49 No3 005-03 JAI & LLS All rights reserved ISSN: 99-8645 wwwatitorg E-ISSN: 87-395 GREY FORECASING AND NEURAL NEWORK MODEL OF SPOR

More information

Grafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space

Grafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space Journal of Machine Learning Research 3 (2003) 1333-1356 Subitted 5/02; Published 3/03 Grafting: Fast, Increental Feature Selection by Gradient Descent in Function Space Sion Perkins Space and Reote Sensing

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

EE5900 Spring Lecture 4 IC interconnect modeling methods Zhuo Feng

EE5900 Spring Lecture 4 IC interconnect modeling methods Zhuo Feng EE59 Spring Parallel LSI AD Algoriths Lecture I interconnect odeling ethods Zhuo Feng. Z. Feng MTU EE59 So far we ve considered only tie doain analyses We ll soon see that it is soeties preferable to odel

More information

Defect-Aware SOC Test Scheduling

Defect-Aware SOC Test Scheduling Defect-Aware SOC Test Scheduling Erik Larsson +, Julien Pouget*, and Zebo Peng + Ebedded Systes Laboratory + LIRMM* Departent of Coputer Science Montpellier 2 University Linköpings universitet CNRS Sweden

More information

Support Vector Machines. Machine Learning Series Jerry Jeychandra Blohm Lab

Support Vector Machines. Machine Learning Series Jerry Jeychandra Blohm Lab Support Vector Machines Machine Learning Series Jerry Jeychandra Bloh Lab Outline Main goal: To understand how support vector achines (SVMs) perfor optial classification for labelled data sets, also a

More information

Efficient Filter Banks And Interpolators

Efficient Filter Banks And Interpolators Efficient Filter Banks And Interpolators A. G. DEMPSTER AND N. P. MURPHY Departent of Electronic Systes University of Westinster 115 New Cavendish St, London W1M 8JS United Kingdo Abstract: - Graphical

More information

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lesson 1 4 October 2017 Outline Learning and Evaluation for Pattern Recognition Notation...2 1. The Pattern Recognition

More information

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words)

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words) 1 A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine (1900 words) Contact: Jerry Farlow Dept of Matheatics Univeristy of Maine Orono, ME 04469 Tel (07) 866-3540 Eail: farlow@ath.uaine.edu

More information

Estimation of ADC Nonlinearities from the Measurement in Input Voltage Intervals

Estimation of ADC Nonlinearities from the Measurement in Input Voltage Intervals Estiation of ADC Nonlinearities fro the Measureent in Input Voltage Intervals M. Godla, L. Michaeli, 3 J. Šaliga, 4 R. Palenčár,,3 Deptartent of Electronics and Multiedia Counications, FEI TU of Košice,

More information

Multi-view Discriminative Manifold Embedding for Pattern Classification

Multi-view Discriminative Manifold Embedding for Pattern Classification Multi-view Discriinative Manifold Ebedding for Pattern Classification X. Wang Departen of Inforation Zhenghzou 450053, China Y. Guo Departent of Digestive Zhengzhou 450053, China Z. Wang Henan University

More information

UNIVERSITY OF TRENTO ON THE USE OF SVM FOR ELECTROMAGNETIC SUBSURFACE SENSING. A. Boni, M. Conci, A. Massa, and S. Piffer.

UNIVERSITY OF TRENTO ON THE USE OF SVM FOR ELECTROMAGNETIC SUBSURFACE SENSING. A. Boni, M. Conci, A. Massa, and S. Piffer. UIVRSITY OF TRTO DIPARTITO DI IGGRIA SCIZA DLL IFORAZIO 3823 Povo Trento (Italy) Via Soarive 4 http://www.disi.unitn.it O TH US OF SV FOR LCTROAGTIC SUBSURFAC SSIG A. Boni. Conci A. assa and S. Piffer

More information

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t. CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when

More information

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair Proceedings of the 6th SEAS International Conference on Siulation, Modelling and Optiization, Lisbon, Portugal, Septeber -4, 006 0 A Siplified Analytical Approach for Efficiency Evaluation of the eaving

More information

3D acoustic wave modeling with a time-space domain dispersion-relation-based Finite-difference scheme

3D acoustic wave modeling with a time-space domain dispersion-relation-based Finite-difference scheme P-8 3D acoustic wave odeling with a tie-space doain dispersion-relation-based Finite-difference schee Yang Liu * and rinal K. Sen State Key Laboratory of Petroleu Resource and Prospecting (China University

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS

EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS Jochen Till, Sebastian Engell, Sebastian Panek, and Olaf Stursberg Process Control Lab (CT-AST), University of Dortund,

More information

Adapting the Pheromone Evaporation Rate in Dynamic Routing Problems

Adapting the Pheromone Evaporation Rate in Dynamic Routing Problems Adapting the Pheroone Evaporation Rate in Dynaic Routing Probles Michalis Mavrovouniotis and Shengxiang Yang School of Coputer Science and Inforatics, De Montfort University The Gateway, Leicester LE1

More information

A Model for the Selection of Internet Service Providers

A Model for the Selection of Internet Service Providers ISSN 0146-4116, Autoatic Control and Coputer Sciences, 2008, Vol. 42, No. 5, pp. 249 254. Allerton Press, Inc., 2008. Original Russian Text I.M. Aliev, 2008, published in Avtoatika i Vychislitel naya Tekhnika,

More information

(6) B NN (x, k) = Tp 2 M 1

(6) B NN (x, k) = Tp 2 M 1 MMAR 26 12th IEEE International Conference on Methods and Models in Autoation and Robotics 28-31 August 26 Międzyzdroje, Poland Synthesis of Sliding Mode Control of Robot with Neural Networ Model Jaub

More information

Algorithms for parallel processor scheduling with distinct due windows and unit-time jobs

Algorithms for parallel processor scheduling with distinct due windows and unit-time jobs BULLETIN OF THE POLISH ACADEMY OF SCIENCES TECHNICAL SCIENCES Vol. 57, No. 3, 2009 Algoriths for parallel processor scheduling with distinct due windows and unit-tie obs A. JANIAK 1, W.A. JANIAK 2, and

More information

ANALYSIS ON RESPONSE OF DYNAMIC SYSTEMS TO PULSE SEQUENCES EXCITATION

ANALYSIS ON RESPONSE OF DYNAMIC SYSTEMS TO PULSE SEQUENCES EXCITATION The 4 th World Conference on Earthquake Engineering October -7, 8, Beijing, China ANALYSIS ON RESPONSE OF DYNAMIC SYSTEMS TO PULSE SEQUENCES EXCITATION S. Li C.H. Zhai L.L. Xie Ph. D. Student, School of

More information

Support Vector Machines MIT Course Notes Cynthia Rudin

Support Vector Machines MIT Course Notes Cynthia Rudin Support Vector Machines MIT 5.097 Course Notes Cynthia Rudin Credit: Ng, Hastie, Tibshirani, Friedan Thanks: Şeyda Ertekin Let s start with soe intuition about argins. The argin of an exaple x i = distance

More information

Research Article Robust ε-support Vector Regression

Research Article Robust ε-support Vector Regression Matheatical Probles in Engineering, Article ID 373571, 5 pages http://dx.doi.org/10.1155/2014/373571 Research Article Robust ε-support Vector Regression Yuan Lv and Zhong Gan School of Mechanical Engineering,

More information

A Note on the Applied Use of MDL Approximations

A Note on the Applied Use of MDL Approximations A Note on the Applied Use of MDL Approxiations Daniel J. Navarro Departent of Psychology Ohio State University Abstract An applied proble is discussed in which two nested psychological odels of retention

More information

Kinematics and dynamics, a computational approach

Kinematics and dynamics, a computational approach Kineatics and dynaics, a coputational approach We begin the discussion of nuerical approaches to echanics with the definition for the velocity r r ( t t) r ( t) v( t) li li or r( t t) r( t) v( t) t for

More information

Computational and Statistical Learning Theory

Computational and Statistical Learning Theory Coputational and Statistical Learning Theory TTIC 31120 Prof. Nati Srebro Lecture 2: PAC Learning and VC Theory I Fro Adversarial Online to Statistical Three reasons to ove fro worst-case deterinistic

More information

A note on the multiplication of sparse matrices

A note on the multiplication of sparse matrices Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani

More information

Domain-Adversarial Neural Networks

Domain-Adversarial Neural Networks Doain-Adversarial Neural Networks Hana Ajakan, Pascal Gerain 2, Hugo Larochelle 3, François Laviolette 2, Mario Marchand 2,2 Départeent d inforatique et de génie logiciel, Université Laval, Québec, Canada

More information

Understanding Machine Learning Solution Manual

Understanding Machine Learning Solution Manual Understanding Machine Learning Solution Manual Written by Alon Gonen Edited by Dana Rubinstein Noveber 17, 2014 2 Gentle Start 1. Given S = ((x i, y i )), define the ultivariate polynoial p S (x) = i []:y

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

CHAPTER 2 LITERATURE SURVEY, POWER SYSTEM PROBLEMS AND NEURAL NETWORK TOPOLOGIES

CHAPTER 2 LITERATURE SURVEY, POWER SYSTEM PROBLEMS AND NEURAL NETWORK TOPOLOGIES 7 CHAPTER LITERATURE SURVEY, POWER SYSTEM PROBLEMS AND NEURAL NETWORK TOPOLOGIES. GENERAL This chapter presents detailed literature survey, briefly discusses the proposed probles, the chosen neural architectures

More information

Warning System of Dangerous Chemical Gas in Factory Based on Wireless Sensor Network

Warning System of Dangerous Chemical Gas in Factory Based on Wireless Sensor Network 565 A publication of CHEMICAL ENGINEERING TRANSACTIONS VOL. 59, 07 Guest Editors: Zhuo Yang, Junie Ba, Jing Pan Copyright 07, AIDIC Servizi S.r.l. ISBN 978-88-95608-49-5; ISSN 83-96 The Italian Association

More information

arxiv: v1 [cs.ds] 17 Mar 2016

arxiv: v1 [cs.ds] 17 Mar 2016 Tight Bounds for Single-Pass Streaing Coplexity of the Set Cover Proble Sepehr Assadi Sanjeev Khanna Yang Li Abstract arxiv:1603.05715v1 [cs.ds] 17 Mar 2016 We resolve the space coplexity of single-pass

More information

Convergence of Hybrid Algorithm with Adaptive Learning Parameter for Multilayer Neural Network

Convergence of Hybrid Algorithm with Adaptive Learning Parameter for Multilayer Neural Network Convergence of Hybrid Algorithm with Adaptive Learning Parameter for Multilayer Neural Network Fadwa DAMAK, Mounir BEN NASR, Mohamed CHTOUROU Department of Electrical Engineering ENIS Sfax, Tunisia {fadwa_damak,

More information

Computational statistics

Computational statistics Computational statistics Lecture 3: Neural networks Thierry Denœux 5 March, 2016 Neural networks A class of learning methods that was developed separately in different fields statistics and artificial

More information

Lecture 12: Ensemble Methods. Introduction. Weighted Majority. Mixture of Experts/Committee. Σ k α k =1. Isabelle Guyon

Lecture 12: Ensemble Methods. Introduction. Weighted Majority. Mixture of Experts/Committee. Σ k α k =1. Isabelle Guyon Lecture 2: Enseble Methods Isabelle Guyon guyoni@inf.ethz.ch Introduction Book Chapter 7 Weighted Majority Mixture of Experts/Coittee Assue K experts f, f 2, f K (base learners) x f (x) Each expert akes

More information

Estimating Average-Case Learning Curves Using Bayesian, Statistical Physics and VC Dimension Methods

Estimating Average-Case Learning Curves Using Bayesian, Statistical Physics and VC Dimension Methods Estiating Average-Case Learning Curves Using Bayesian, Statistical Physics and VC Diension Methods David Haussler University of California Santa Cruz, California Manfred Opper Institut fur Theoretische

More information

Variations on Backpropagation

Variations on Backpropagation 2 Variations on Backpropagation 2 Variations Heuristic Modifications Moentu Variable Learning Rate Standard Nuerical Optiization Conjugate Gradient Newton s Method (Levenberg-Marquardt) 2 2 Perforance

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement

More information

MSEC MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL SOLUTION FOR MAINTENANCE AND PERFORMANCE

MSEC MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL SOLUTION FOR MAINTENANCE AND PERFORMANCE Proceeding of the ASME 9 International Manufacturing Science and Engineering Conference MSEC9 October 4-7, 9, West Lafayette, Indiana, USA MSEC9-8466 MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL

More information

ACTIVE VIBRATION CONTROL FOR STRUCTURE HAVING NON- LINEAR BEHAVIOR UNDER EARTHQUAKE EXCITATION

ACTIVE VIBRATION CONTROL FOR STRUCTURE HAVING NON- LINEAR BEHAVIOR UNDER EARTHQUAKE EXCITATION International onference on Earthquae Engineering and Disaster itigation, Jaarta, April 14-15, 8 ATIVE VIBRATION ONTROL FOR TRUTURE HAVING NON- LINEAR BEHAVIOR UNDER EARTHQUAE EXITATION Herlien D. etio

More information

Gary J. Balas Aerospace Engineering and Mechanics, University of Minnesota, Minneapolis, MN USA

Gary J. Balas Aerospace Engineering and Mechanics, University of Minnesota, Minneapolis, MN USA μ-synthesis Gary J. Balas Aerospace Engineering and Mechanics, University of Minnesota, Minneapolis, MN 55455 USA Keywords: Robust control, ultivariable control, linear fractional transforation (LFT),

More information

DISSIMILARITY MEASURES FOR ICA-BASED SOURCE NUMBER ESTIMATION. Seungchul Lee 2 2. University of Michigan. Ann Arbor, MI, USA.

DISSIMILARITY MEASURES FOR ICA-BASED SOURCE NUMBER ESTIMATION. Seungchul Lee 2 2. University of Michigan. Ann Arbor, MI, USA. Proceedings of the ASME International Manufacturing Science and Engineering Conference MSEC June -8,, Notre Dae, Indiana, USA MSEC-7 DISSIMILARIY MEASURES FOR ICA-BASED SOURCE NUMBER ESIMAION Wei Cheng,

More information

Machine Learning Basics: Estimators, Bias and Variance

Machine Learning Basics: Estimators, Bias and Variance Machine Learning Basics: Estiators, Bias and Variance Sargur N. srihari@cedar.buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics in Basics

More information

Easy Evaluation Method of Self-Compactability of Self-Compacting Concrete

Easy Evaluation Method of Self-Compactability of Self-Compacting Concrete Easy Evaluation Method of Self-Copactability of Self-Copacting Concrete Masanori Maruoka 1 Hiroi Fujiwara 2 Erika Ogura 3 Nobu Watanabe 4 T 11 ABSTRACT The use of self-copacting concrete (SCC) in construction

More information

arxiv: v1 [cs.lg] 8 Jan 2019

arxiv: v1 [cs.lg] 8 Jan 2019 Data Masking with Privacy Guarantees Anh T. Pha Oregon State University phatheanhbka@gail.co Shalini Ghosh Sasung Research shalini.ghosh@gail.co Vinod Yegneswaran SRI international vinod@csl.sri.co arxiv:90.085v

More information

Graphical Models in Local, Asymmetric Multi-Agent Markov Decision Processes

Graphical Models in Local, Asymmetric Multi-Agent Markov Decision Processes Graphical Models in Local, Asyetric Multi-Agent Markov Decision Processes Ditri Dolgov and Edund Durfee Departent of Electrical Engineering and Coputer Science University of Michigan Ann Arbor, MI 48109

More information

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians Using EM To Estiate A Probablity Density With A Mixture Of Gaussians Aaron A. D Souza adsouza@usc.edu Introduction The proble we are trying to address in this note is siple. Given a set of data points

More information

TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES

TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES S. E. Ahed, R. J. Tokins and A. I. Volodin Departent of Matheatics and Statistics University of Regina Regina,

More information

Predictive Vaccinology: Optimisation of Predictions Using Support Vector Machine Classifiers

Predictive Vaccinology: Optimisation of Predictions Using Support Vector Machine Classifiers Predictive Vaccinology: Optiisation of Predictions Using Support Vector Machine Classifiers Ivana Bozic,2, Guang Lan Zhang 2,3, and Vladiir Brusic 2,4 Faculty of Matheatics, University of Belgrade, Belgrade,

More information

Forecasting Financial Indices: The Baltic Dry Indices

Forecasting Financial Indices: The Baltic Dry Indices International Journal of Maritie, Trade & Econoic Issues pp. 109-130 Volue I, Issue (1), 2013 Forecasting Financial Indices: The Baltic Dry Indices Eleftherios I. Thalassinos 1, Mike P. Hanias 2, Panayiotis

More information

Comparison of Stability of Selected Numerical Methods for Solving Stiff Semi- Linear Differential Equations

Comparison of Stability of Selected Numerical Methods for Solving Stiff Semi- Linear Differential Equations International Journal of Applied Science and Technology Vol. 7, No. 3, Septeber 217 Coparison of Stability of Selected Nuerical Methods for Solving Stiff Sei- Linear Differential Equations Kwaku Darkwah

More information

1 Bounding the Margin

1 Bounding the Margin COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost

More information

Reduced Length Checking Sequences

Reduced Length Checking Sequences Reduced Length Checing Sequences Robert M. Hierons 1 and Hasan Ural 2 1 Departent of Inforation Systes and Coputing, Brunel University, Middlesex, UB8 3PH, United Kingdo 2 School of Inforation echnology

More information

Optimal Control of Nonlinear Systems Using the Shifted Legendre Polynomials

Optimal Control of Nonlinear Systems Using the Shifted Legendre Polynomials Optial Control of Nonlinear Systes Using Shifted Legendre Polynoi Rahan Hajohaadi 1, Mohaad Ali Vali 2, Mahoud Saavat 3 1- Departent of Electrical Engineering, Shahid Bahonar University of Keran, Keran,

More information

Randomized Recovery for Boolean Compressed Sensing

Randomized Recovery for Boolean Compressed Sensing Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch

More information

Short Papers. Test Data Compression and Decompression Based on Internal Scan Chains and Golomb Coding

Short Papers. Test Data Compression and Decompression Based on Internal Scan Chains and Golomb Coding IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, VOL. 1, NO. 6, JUNE 00 715 Short Papers Test Data Copression and Decopression Based on Internal Scan Chains and Golob Coding

More information

On Constant Power Water-filling

On Constant Power Water-filling On Constant Power Water-filling Wei Yu and John M. Cioffi Electrical Engineering Departent Stanford University, Stanford, CA94305, U.S.A. eails: {weiyu,cioffi}@stanford.edu Abstract This paper derives

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,

More information

1 Generalization bounds based on Rademacher complexity

1 Generalization bounds based on Rademacher complexity COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #0 Scribe: Suqi Liu March 07, 08 Last tie we started proving this very general result about how quickly the epirical average converges

More information