THE PERFORMANCE OF CLASSIFICATION RULES PRODUCED BY SYMBOLIC INDUCTIVE METHODS

Size: px
Start display at page:

Download "THE PERFORMANCE OF CLASSIFICATION RULES PRODUCED BY SYMBOLIC INDUCTIVE METHODS"

Transcription

1 Six:h inlt!rnajiona/ Symposium on.~fe/hod%gies jor ifllelligefict! SYSlt!ms Char/oue,.vC.. OClober 16 19, 1991 ~/-t5 '7/-~ USI~G GE~ETIC ALGORITHi\1S TO IMPROVE THE PERFORMANCE OF CLASSIFICATION RULES PRODUCED BY SYMBOLIC INDUCTIVE METHODS Jerzy Bala Kenneth Dejong Peter Pachowicz Center for Artificial Intelligence, George Mason University 4400 University Drive, Fairfax, VA Abstract: In this paper we present a novel way 0/ combining symbolic inductive methods and genetic algorithms (GAs) applied to produce high-performance classification rules. The presented method consists 0/ two phases. In the first one the algorithm induvtively learns a set of classification rules from noisy input examples. In the second phase the worst performing rule is optimized by GAs techniques. Experimental results are presented/or twelve classes o/noisy data obtained/rom textured images. 1 Introduction One fundamental weakness of inductive learning (except for special cases) is the fact that the acquired knowledge cannot be validated. Traditional inquiries into inductive inference have therefore dealt with questions of what are the best criteria for guiding the selection of inductive assertions, and how these assertiops can be confirmed. The goal of inference is to formulate plausible general assertions that explain the given facts and are able to predict new facts. For a given set of facts, a potentially infinite number of hypotheses can be generated that imply these facts. A preference criterion is therefore necessary to provide constraints and reduce the infinite choice to one hypothesis or a few more preferable ones. A typical way of defining such criterion is to specify the preferable propenies of the hypothesis, for example, to require that the hypothesis is the shonest or the most economical description consistent with all the facts. Even if the preferred criterion is defined it can generate other problems: e.g. performance on the future data. This is especially imponant if the initial data has some distribution of noisy and irrelevant examples. We propose a novel hybrid approach in which, frrst, a classic rule induction system (AQ) produces shon, simple, complete and, consistent rules; and then GAs are used to improve the performance of the rule via an evolutionary mechanism. By dropping the requirement for a rule

2 to be complete and consistent with respect to the initial learning dara, we reduce the coverage of noisy and irrelevant components in the rule structure. Since the AQ learning algorithm is the induction learning method in our method the next section presents a short overview of this learning method. Section 3 presents the GA phase of our method and the experimental results are presented in section 4. 2 AQ learning algorithm The AQ algorithm [Michalski et al., 1986] learns atuibutional description from examples. When building a decision rule. AQ performs a heuristic search through a space of logical expressions to determine those that account for all positive examples and no negative examples. Because there are usually many such complete and consistent expressions, the goal of AQ is to find the most preferred one, according to some criterion. Learning examples are given in the form of events. which are vectors of attribute values. Events represent different decision classes. For each class a decision rule is produced that covers all examples of that class and none of other classes. The concept descriptions learned by AQ algorithm are represented in VL 1, the Variable Valued Logic System 1 [Michalski. 1972]. and are used to represent attributional concept descriptions. A description of a concept is a disjunctive normal form which is called a cover. A cover is a disjunction of complexes. A complex is a conjunction of selectors. A selector is a form: [L#R] (1) where. L is called the referee. which is an attribute. R is called the referent, which is a set of values in the domain of the attribute L. # is one of the following relational symbols: =, <, >. >=, <=, <>. The following is an example of the five selectors AQ complex (equality is used as a relational symbol): [xl=1.. 3][x2=1][x4=O][x6=1..7][x8=1] Dots in the description represent a range of possible values of a given attribute. An example of simple, four complexes. AQ description is presented below: 1 2 [xl=7..8] [x2=8.. 19] [x ] [x5=4..54] [x6=o..3] [xl= ] [x3=11..14] [x4= ] [x6=o.. 9] [x7=o [xl=9.. 18] [x3=16..21] [x4=9.. 10] [xl=1o.. 14] [x3= ] [x4= ] [x7=4.. 5] 2

3 3 Genetic Algorithm's Phase The rules generated by the AQ algorithm have the propeny of being complete and consistent with respect to the learning examples. These properties mean that each class covers its learning examples (positive examples) and does not cover any negative examples (negative examples are all examples belonging to other classes), \\'hen the AQ algorithm is applied to finergrained and noisy symbolic representations the consistency and completeness conditions overconstrain the generated class description. This, in turn, leads to poorer predictive accuracy on future data. Conversely, by relaxing the requirement of consistency and completeness, predictive accuracy in noisy domains can be improved. In the proposed method we use GAs to improve the performance of initially consistent and complete rules via the evolutionary mechanism. A genetic algorithm [De Jong. 1988] maintains a constant-sized population of candidate solutions. known as individuals. At each iteration, each individual is evaluated and recombined with others on the basis of its overall quality or fitness. The expected number of times an individual is selected for recombination is proportional to its fitness relative to the rest of population. The power of a genetic algorithm lies in its ability to exploit, in a highly efficient manner, information about a large number of individuals. By allocating more reproductive occurrences to above average individuals, the overall affect is an increase of the population's average fitness. New individuals are created using two main genetic recombination operators known as crossover and mutation. Crossover operates by selecting a random location in the genetic saing of the parents (crossover point) and concatenating the initial segments of one parent with the final segment parent to create a new child. A second child is simultaneously generated using the remaining segments of the two parents. Mutation provides for occasional disturbances in the crossover operation to insure diversity in the genetic strings over long periods of rime and to prevent stagnation in the convergence of the optimization's technique. In order to apply GAs. one needs to choose a representation, defme the operators and the performance measure. In traditional GA's individuals in the population are typically represented using a binary string notation to promote efficiency and application independence in the genetic operations. The mathematical analysis of GAs shows that they work best when the internal representation encourages the emergence of useful building blocks that can be subsequently combined with others to produce improved performance and string representations are just one of many ways of achieving this. In old' application of GAs we do not use string representations. Each individual of the population is represented in VLl (previous section). as the different version of a given rule that is a subject for the optimization. The VL 1 representation is very "natural" for the optimization problem defined in this paper (search for a better performing rule) and can easily be manipulated by GA operators. The next section describes GA operators chosen for our method. 3

4 3.1 GA operators An initial population of rules is created by causing small variations to an existing rule. Then, a population of different variations of this rule is evolved by using GA's operators. Each rule in the population is evaluated based on its peiiormance within the set of initial rules. To evaluate the rule we use the tuning data which are pan of the set of learning examples. The perfonnance of the best rule in the population is monitored by matching this rule with the testing examples. Mutation is performed by introducing small changes to the condition pans (selectors) of a selected disjunct (rule complex) of a given rule. Th,e selectors to be changed are chosen by randomly generating two pointers. the first one that points to the rule disjunct complex. and the second one points to the selector of this disjunct. Random changes are introduced to the left-most or right-most (this is also chosen randomly) value of this selector. For example, the selector [x5= ] can be changed to one of the following; [x5=3,io [x5=3,io..241, [x5= ] or [x5= ]. Such a mutation process samples the space of possible boundaries between rules to minimize the coverage of noisy and irrelevant components of a given rule. The crossover operation is performed by splitting rule disjuncts into two parts, upper disjuncts and lower disjuncts. These parts are exchanged between parent rules to produce new child rules. Since degree of match of a given instance depends on the degree of match of this instance to each disjunct of that rule, this exchange process enables inheritance of information about strong disjuncts (suongly matching) in the individuals of the next evolved population. An example ofcrossover applied to short, four disjuncts rules is depicted below: Parent rule I I [xl=7..8] [x2=8.. 19] [x3=8.. 13] [xs=4..54] [x6=o..3] 2 [xl=is.. 54] [x3=11..14] [x4= ] [x6=o.. 9] [x7=o..11] crossover position [xl=9..18] [x3=16..21] [x4=9..10] 4 [xl=io.. 14] [x3= ] [x4=14..54] [x7=4.. 5] Parent rule 2 I [x3= ] [x4= ] [xs=o.. 6] [x7=5.. 12] 2 [xl=8.. 25] [x3=8.. 13] [x4=9.. 11] [x5=o.. 3] crossover position [x4=o. 22] [x5=8.. 9] [x6=o.. 7] [x7=11..48] 4 [x2=s. 8] [x3=7.. 8] [x4=8.. 11] [x5=o.. 3] Result of the crossover operation (one of two child rules) I [xl=7. 8] [x2=8.. 19] [x3=8.. 13] [xs=4..54] [x6=o..3] 2 [xl=is.. 54] [x3=11..14] [x4= ] [x6=o..9] [x7=o.. 11] 3 [x4=o.. 22] [x5=8.. 9] [x6=o.. 7] [x7=11..48] 4 [x2=5.. 8] [x3=7.. 8] [x4=8.. 11] [xs=o. 3] 4

5 The GA operators as described above are very effective in situations where no a priori infonnation is given on the distribution of attribute values. Son-nonnal distribution of attribute values complicates the AQ generated descriptions, especially if the learned rules consistently and completely cover input instances of the data. Graph 1 presents the most representative samples of non-normal attribute distribution obtained from the experiment with textured image classes described in section 4. Class: C9; Attribute:,,5 8~--~----,---~----,---~~ Class: C8; Attribute: xl O;--'~~~-r----r-~-+~~~ O~~--~---+--P-~~--+--r-;~ o 10 AUribute value o 10 AUribute value Class: C8; Attribute: x3 ~.g -l t-_,r--ii--1t a ~ -+--I-----t'--+--tfo~ + t i+---~--~~~--+---~~--~~ l+---~~--r---~--~~r-~ Class: Cll; Attribute:...7 4,-----~--~----~----~--~~ O~~--+--p--P-~~ ~~~ o 10 AUribuae value O~~~--~~~~--~~-r~~ o 10 AUribute value Graph 1. Examples of non-nonnal attribute distribution. The solid line corresponds to smoothed distribution of an attribute and the dotted line corresponds to the approximated nonnal distribution. The left hand diagrams of Graph 1 present a relatively simple attribute distribution. However, the right hand diagrams of Graph I present two cases of more complex attribute distributions. The upper right diagram indicates the possible formation of more than one cluster of training data. The lower right diagram presents very complex distribution of an attribute without distinction of regular clusters. One can notice that x7 5

6 attribute does not carry significant information about CII class. Considering the above complexity of the attributes. we postulate that this domain (data in a representation space expressed by such attributes) is a good candidate for the GA type of searches, where random changes (mutation) and some directional sampling (crossover) can yield high performance solutions (optimized rules). 3.2 Rule performance evaluation Each rule candidate in the population is evaluated by plugging it into the set of other rules and calculating the confusion mattix for the tuning examples. The confusion matrix represents infonnation on correct and incorrect classification results. This matrix is obtained by calculating a degree of flexible match between a tuning instance and a given class description. The row entries it) the matrix represent a percentage of matched instances from a given class (row index) to all class descriptions (column index). The following is an example of the confusion matrix for 12 classes: Cl C2 C3 C4 C5 C6 Cl C8 C9 CIO Cl1 C12 Cl 45 S C2 S C S C C C C C C I 2 CI0 0 0 I Cll C Average recognition=74.67% Average mis-classitication=4.83 The above confusion matrix represents classification results on data extracted from textured images. The method of extraction and texture classification by inductive learning is not the subject of research presented in this paper. The reader can find various information on the method of learning texture descriptions in the following papers [Bala and Pachowicz. 1991]. Let us suppose that the class Cll is the one chosen to be evolved by GAs mechanism. First we mutate this class representation multiple times to produce an initial population of different individuals of that class. To test the performance of a given individual we calculate the confusion matrix. The confusion matrix shows us how this specific variation of class Cll 6

7 perfonns when tested with other classes using the set of tuning examples. For example. we can see from me above confusion matrix mat the class C 11 has me following performance: Cl C2 C3 C4 C5 C6 C7 C8 C9 ClO Cll Cl2 CII The above values represent percentage of instances of class C 11 (set of tuning instances) matched to learned descriptions of all 12 classes. Ifclass CII is to be optimized. by GAs we have to evaluate the performance of each individual from the population of that class by calculating the performance measure of different candidates of that class with respect to other classes. Thus, as the performance evaluation measure of each individual of the population we use the ratio of correct recognition rates to incorrect recognition rates; CC/MC, where CC is an average recognition for correctly recognized classes (an average of entries on the diagonal of the confusion matrix). and MC is an average mis-classification (an average of entries outside the diagonal of the confusion matrix). For the above confusion matrix CC/MC=74.67/4.83= The next section describes how the distance measure between an example (tuning or testing) and a rule is calculated Reeo&nizin& class membership throu&h nexible matehin& Flexible matching measures the degree of closeness between an instance and the conditional pan of a rule. Such closeness is computed as distance value between an instance and rule complex within the attribute space. This distance measure takes any value in the range from o (Le., does not match) to 1 (Le. matches). Calculation of this distance measure for a single test instance is executed. according to the following schema. For a given condition ofa rule: [ Xn =valjl and an instance where Xn = valk ' the normalized distance measure is 1 ( I val j val k II #levels ) (2) where #levels is the total number of attrib\1te values. The confidence level of a rule is computed by multiplying evaluation values of each condition of the rule. The total evaluation of class membership of a given test instance is equal to the confidence level of the best matching rule, i.e. the rule of the highest confidence value. For example. the confidence value c for matching the following complex of a rule [ xl =0 ] [ x7 =10 ] [ x8 = ] with a given instance x = < 4. S. 24, 34, O. 12, 6, 25 > #levels=55 is computed as follows: and the number of attribute values c x } = 1 ( 10-41/55) =.928 C x 7 = 1 - ( / 55) =.928 c x 8 = 1 - ( ) =.91 Cx2, cx3. cx4' cxs, cx6 = 1 7

8 c = cxl * Cx2 * Cx) * Cx4 * Cx5 * Cx6 * Cx7 * C x 8 = 0.78 The recognition process yields class membership for which the confidence level is the highest among matched rules. Calculated confidence level, however, is not a probability measure and it can yield more than one class membership. It means that for a given preclassified test dataset. the system recognition effectiveness is calculated as the number of correctly classified instances to the total number of instances of the dataset. The above described method of rule evaluation by flexible matching and confidence level calculation has imponant consequences for the crossover operation. Since the degree of match of an instance to a rule depends on the confidence level of this instance to each disjunct of that rule, the swapping process introduced by crossover operator enables inheritance of infonnation represented in the strong disjuncts (with high confidence level) of rule candidates through generations of a population. 4. Experimental results In the experiment we used 12 classes of texture data. The initial descriptions were generated by the AQ module using 200 examples per class. Another set of 100 examples was used to calculat!' the perfonnance measure for the GAs cycle. The testing set had 200 examples extracted from different image areas other than training and tuning data.. Examples of a given class were constructed using 8 attributes and an extraction process based on Law's mask [Bala and Pachowicz, 1991]. The "weakest" class (with the lowest matching results with testing data) chosen for the experiment was class Cll (like in the confusion matrix example in section 3.2) with 40 disjuncts. Graph 2 represents results of the experiment. White circles of the diagrams represent characteristics obtained for tuning data used to guide the genetic search. Characteristics mapped by black circles were obtained for testing data. Upper diagram monitors the performance of genetically evolved description of Cll texture. When all 300 examples were used to generate rules, the average classification rate for this class (when tested with 200 examples) was below 45%. When the set of 300 examples was split into two parts, 200 for the initial inductive learning and 100 as the tuning data for GAs cycle, the correct classification rates obtainedtin the 30th evolution was above 60%. That is a significant increase in comparison with 45% obtained from inductive learning only. The bottom diagrams represent the evaluation function used by genetic algorithm (COMe, section 3.2) in order to guide the genetic sean:h. The evaluation function was calculated as a rate of the correct classifications to mis-classifications for all twelve texture classes. These diagrams are depicted for both testing and tuning data. The increases of CC/MC on both diagrams represent an overall improvement of system recognition performance. The system performance was investigated for a larger number of GA generation steps. However, it appears, that the substantial increase was reached both for the CII class description and for the overall system performance in a very few generation steps (Le., in 10 steps). 8

9 o Results for tuning dat.a Results for testing data 30~----~A ~ 20;---~--~----~--~~--~--~ o 10 Generations ~ 0 "1 ~ r J roo II MM be, u ~ /\.J ~ f \-\ o 10 Generations Generations 30 Graph 2. Results of experiment with 12 classes. s. Conclusion This paper presents a novel approach to the optimization of the classification rule descriptions through GAs techniques. We recognize rule optimization as an emerging research area in the creation of autonomous intelligent systems. Two issues appear to be important for this research. First, rule optimization is absolutely necessary in such problems where the system must be protected against the accumulation of noisy components and where attributes used to describe initial data have complex. non-nonna! distributions. The method can be augmented by 9

10 optimizing input data formation (searching for the best subset of attributes) prior to rule generation [Vafaie, De Jong 1991]. Secondly, for inducti...ely generated rules (falsity preserving rules) there should be some method of validating these rules (or choosing the most preferred inductive rule) before applying them to future data. The methoo presented in this paper tries to offer some solutions to these problems. The presented experiments on applying this hybrid methoo to finer-grained symbolic representations of vision data are encouraging. Highly disjunctive descriptions obtained by inductive learning from this data were easily manipulated by chosen GA operators and a substantial increase in the overall system perfonnance was observed in a few generations. Although there exists substantial differences between GAs and symbolic induction methods (learning algorithm, performance elements, knowledge representation), this method serves as a promising example that a better understanding of abilities of each approach (identifying common building blocks) can lead to novel and useful ways of combining them. This paper concentrates primarily on the proposed methodology. More experiments are needed to draw concrete conclusions. Future research will attempt to thoroughly test this approach on numerous data sets. We will also examine more thoroughly mutation and crossover operators. Acknowledgement: This research was supponed in pan by the Office of Naval Research under grants No. NOOOI4-88-K-0397 and No. NOOO14-88-K-0226, and in pan by the Defense Advan~ed Research Projects Agency under the grant administered by the Office of Naval Research No. NOOOI4-K The authors wish to thank: Janet Holmes and Haleh Vafaie for editing suggestions. References Bala. I.W. and Pachowicz, P.W., "Application of Symbolic Machine Learning to the Recognition oftexture Concepts" 7th IEEE Conference on Artificial Intelligence Applications, Miami Beach FL, February De Jong, K "Learning with Genetic Algorithms: An Overview". Machine Learning vol. 3; , Kluwer Academic Publishers Vafaie H.and De Jong, K.,"Improving the Performance of a Rule Induction System Using Genetic Algorithms"; in preparation.(center for AI, GMU 1991) Fitzpatrick. J.M. and Grefenstette I., "Genetic Algorithms in Noisy Environments", Machine Leaming 3 pp (1988). MichalskiR. S. "AQVAUI--ComputerImplementation ofa Variable-Valued Logic System VLl and Examples of Its Application to Pattern Recognition", First International Joint Conference on Pattern Recognition, October 30, 1973, Washington D.C. Michalski, R.S. Mozetic I., Hong I.R., Lavrac N. "The AQ15 Inductive Learning System", Report No. UIUCDCS-R , Department of Computer Science, University of Illinois at Urbane-Champaign, July

Learning From Inconsistent and Noisy Data: The AQ18 Approach *

Learning From Inconsistent and Noisy Data: The AQ18 Approach * Eleventh International Symposium on Methodologies for Intelligent Systems, Warsaw, pp. 411-419, 1999 Learning From Inconsistent and Noisy Data: The AQ18 Approach * Kenneth A. Kaufman and Ryszard S. Michalski*

More information

Genetic Algorithms: Basic Principles and Applications

Genetic Algorithms: Basic Principles and Applications Genetic Algorithms: Basic Principles and Applications C. A. MURTHY MACHINE INTELLIGENCE UNIT INDIAN STATISTICAL INSTITUTE 203, B.T.ROAD KOLKATA-700108 e-mail: murthy@isical.ac.in Genetic algorithms (GAs)

More information

Looking Under the EA Hood with Price s Equation

Looking Under the EA Hood with Price s Equation Looking Under the EA Hood with Price s Equation Jeffrey K. Bassett 1, Mitchell A. Potter 2, and Kenneth A. De Jong 1 1 George Mason University, Fairfax, VA 22030 {jbassett, kdejong}@cs.gmu.edu 2 Naval

More information

Lecture 9 Evolutionary Computation: Genetic algorithms

Lecture 9 Evolutionary Computation: Genetic algorithms Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Simulation of natural evolution Genetic algorithms Case study: maintenance scheduling with genetic

More information

Lecture 22. Introduction to Genetic Algorithms

Lecture 22. Introduction to Genetic Algorithms Lecture 22 Introduction to Genetic Algorithms Thursday 14 November 2002 William H. Hsu, KSU http://www.kddresearch.org http://www.cis.ksu.edu/~bhsu Readings: Sections 9.1-9.4, Mitchell Chapter 1, Sections

More information

CSC 4510 Machine Learning

CSC 4510 Machine Learning 10: Gene(c Algorithms CSC 4510 Machine Learning Dr. Mary Angela Papalaskari Department of CompuBng Sciences Villanova University Course website: www.csc.villanova.edu/~map/4510/ Slides of this presenta(on

More information

Search. Search is a key component of intelligent problem solving. Get closer to the goal if time is not enough

Search. Search is a key component of intelligent problem solving. Get closer to the goal if time is not enough Search Search is a key component of intelligent problem solving Search can be used to Find a desired goal if time allows Get closer to the goal if time is not enough section 11 page 1 The size of the search

More information

Artificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence

Artificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence Artificial Intelligence (AI) Artificial Intelligence AI is an attempt to reproduce intelligent reasoning using machines * * H. M. Cartwright, Applications of Artificial Intelligence in Chemistry, 1993,

More information

Choosing Variables with a Genetic Algorithm for Econometric models based on Neural Networks learning and adaptation.

Choosing Variables with a Genetic Algorithm for Econometric models based on Neural Networks learning and adaptation. Choosing Variables with a Genetic Algorithm for Econometric models based on Neural Networks learning and adaptation. Daniel Ramírez A., Israel Truijillo E. LINDA LAB, Computer Department, UNAM Facultad

More information

An Evolutionary Programming Based Algorithm for HMM training

An Evolutionary Programming Based Algorithm for HMM training An Evolutionary Programming Based Algorithm for HMM training Ewa Figielska,Wlodzimierz Kasprzak Institute of Control and Computation Engineering, Warsaw University of Technology ul. Nowowiejska 15/19,

More information

Selected Algorithms of Machine Learning from Examples

Selected Algorithms of Machine Learning from Examples Fundamenta Informaticae 18 (1993), 193 207 Selected Algorithms of Machine Learning from Examples Jerzy W. GRZYMALA-BUSSE Department of Computer Science, University of Kansas Lawrence, KS 66045, U. S. A.

More information

Decision Trees (Cont.)

Decision Trees (Cont.) Decision Trees (Cont.) R&N Chapter 18.2,18.3 Side example with discrete (categorical) attributes: Predicting age (3 values: less than 30, 30-45, more than 45 yrs old) from census data. Attributes (split

More information

Efficiently merging symbolic rules into integrated rules

Efficiently merging symbolic rules into integrated rules Efficiently merging symbolic rules into integrated rules Jim Prentzas a, Ioannis Hatzilygeroudis b a Democritus University of Thrace, School of Education Sciences Department of Education Sciences in Pre-School

More information

DETECTION OF ATMOSPHERIC RELEASES

DETECTION OF ATMOSPHERIC RELEASES J1. MACHINE LEARNING EVOLUTION FOR THE SOURCE DETECTION OF ATMOSPHERIC RELEASES G. Cervone 1 and P. Franzese 2 1 Dept. of Geography and Geoinformation Science, George Mason University 2 Center for Earth

More information

TUNING ROUGH CONTROLLERS BY GENETIC ALGORITHMS

TUNING ROUGH CONTROLLERS BY GENETIC ALGORITHMS TUNING ROUGH CONTROLLERS BY GENETIC ALGORITHMS Teresa Chiu Department of Mathematics and Computer Science Sun Jose State University San Jose, California 95192 T.Y. Lin* Department of Electrical and Computer

More information

Representation and Hidden Bias II: Eliminating Defining Length Bias in Genetic Search via Shuffle Crossover

Representation and Hidden Bias II: Eliminating Defining Length Bias in Genetic Search via Shuffle Crossover Representation and Hidden Bias II: Eliminating Defining Length Bias in Genetic Search via Shuffle Crossover Abstract The traditional crossover operator used in genetic search exhibits a position-dependent

More information

An Evolution Strategy for the Induction of Fuzzy Finite-state Automata

An Evolution Strategy for the Induction of Fuzzy Finite-state Automata Journal of Mathematics and Statistics 2 (2): 386-390, 2006 ISSN 1549-3644 Science Publications, 2006 An Evolution Strategy for the Induction of Fuzzy Finite-state Automata 1,2 Mozhiwen and 1 Wanmin 1 College

More information

A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997

A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997 A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997 Vasant Honavar Artificial Intelligence Research Laboratory Department of Computer Science

More information

AN EVOLUTIONARY ALGORITHM TO ESTIMATE UNKNOWN HEAT FLUX IN A ONE- DIMENSIONAL INVERSE HEAT CONDUCTION PROBLEM

AN EVOLUTIONARY ALGORITHM TO ESTIMATE UNKNOWN HEAT FLUX IN A ONE- DIMENSIONAL INVERSE HEAT CONDUCTION PROBLEM Proceedings of the 5 th International Conference on Inverse Problems in Engineering: Theory and Practice, Cambridge, UK, 11-15 th July 005. AN EVOLUTIONARY ALGORITHM TO ESTIMATE UNKNOWN HEAT FLUX IN A

More information

Predicting Dynamical Systems

Predicting Dynamical Systems two projects in which a genetic algorithm is used to solve such search problems-one of predicting dynamical systems, and the other of predicting the structure of proteins. Predicting Dynamical Systems

More information

Data Warehousing & Data Mining

Data Warehousing & Data Mining 13. Meta-Algorithms for Classification Data Warehousing & Data Mining Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 13.

More information

Optimization Models for Detection of Patterns in Data

Optimization Models for Detection of Patterns in Data Optimization Models for Detection of Patterns in Data Igor Masich, Lev Kazakovtsev, and Alena Stupina Siberian State University of Science and Technology, Krasnoyarsk, Russia is.masich@gmail.com Abstract.

More information

A. Pelliccioni (*), R. Cotroneo (*), F. Pungì (*) (*)ISPESL-DIPIA, Via Fontana Candida 1, 00040, Monteporzio Catone (RM), Italy.

A. Pelliccioni (*), R. Cotroneo (*), F. Pungì (*) (*)ISPESL-DIPIA, Via Fontana Candida 1, 00040, Monteporzio Catone (RM), Italy. Application of Neural Net Models to classify and to forecast the observed precipitation type at the ground using the Artificial Intelligence Competition data set. A. Pelliccioni (*), R. Cotroneo (*), F.

More information

GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS

GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS A genetic algorithm is a random search technique for global optimisation in a complex search space. It was originally inspired by an

More information

Section on Survey Research Methods JSM 2010 STATISTICAL GRAPHICS OF PEARSON RESIDUALS IN SURVEY LOGISTIC REGRESSION DIAGNOSIS

Section on Survey Research Methods JSM 2010 STATISTICAL GRAPHICS OF PEARSON RESIDUALS IN SURVEY LOGISTIC REGRESSION DIAGNOSIS STATISTICAL GRAPHICS OF PEARSON RESIDUALS IN SURVEY LOGISTIC REGRESSION DIAGNOSIS Stanley Weng, National Agricultural Statistics Service, U.S. Department of Agriculture 3251 Old Lee Hwy, Fairfax, VA 22030,

More information

The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan

The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan ABSTRACT The genetic algorithm is an adaptive search method that has the ability for a smart search

More information

Hypothesis Evaluation

Hypothesis Evaluation Hypothesis Evaluation Machine Learning Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Hypothesis Evaluation Fall 1395 1 / 31 Table of contents 1 Introduction

More information

CS 6375 Machine Learning

CS 6375 Machine Learning CS 6375 Machine Learning Decision Trees Instructor: Yang Liu 1 Supervised Classifier X 1 X 2. X M Ref class label 2 1 Three variables: Attribute 1: Hair = {blond, dark} Attribute 2: Height = {tall, short}

More information

Lecture 2: Foundations of Concept Learning

Lecture 2: Foundations of Concept Learning Lecture 2: Foundations of Concept Learning Cognitive Systems II - Machine Learning WS 2005/2006 Part I: Basic Approaches to Concept Learning Version Space, Candidate Elimination, Inductive Bias Lecture

More information

Metaheuristics and Local Search

Metaheuristics and Local Search Metaheuristics and Local Search 8000 Discrete optimization problems Variables x 1,..., x n. Variable domains D 1,..., D n, with D j Z. Constraints C 1,..., C m, with C i D 1 D n. Objective function f :

More information

[Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4]

[Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4] 1 EVOLUTIONARY ALGORITHMS [Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4] Evolutionary computation Prototypical GA An example: GABIL Schema theorem Genetic Programming Individual learning and population

More information

Modeling of a Winding Machine Using Genetic Programming

Modeling of a Winding Machine Using Genetic Programming Modeling of a Winding Machine Using Genetic Programming Abo El-Abbass Hussian +, Alaa Sheta +, Mahmoud Kamel ++ Mohammed Telbaney + and Ashraf Abdelwahab + + Computers and Systems Department, Electronics

More information

Binary Particle Swarm Optimization with Crossover Operation for Discrete Optimization

Binary Particle Swarm Optimization with Crossover Operation for Discrete Optimization Binary Particle Swarm Optimization with Crossover Operation for Discrete Optimization Deepak Singh Raipur Institute of Technology Raipur, India Vikas Singh ABV- Indian Institute of Information Technology

More information

Crossover Techniques in GAs

Crossover Techniques in GAs Crossover Techniques in GAs Debasis Samanta Indian Institute of Technology Kharagpur dsamanta@iitkgp.ac.in 16.03.2018 Debasis Samanta (IIT Kharagpur) Soft Computing Applications 16.03.2018 1 / 1 Important

More information

Genetic Algorithms. Seth Bacon. 4/25/2005 Seth Bacon 1

Genetic Algorithms. Seth Bacon. 4/25/2005 Seth Bacon 1 Genetic Algorithms Seth Bacon 4/25/2005 Seth Bacon 1 What are Genetic Algorithms Search algorithm based on selection and genetics Manipulate a population of candidate solutions to find a good solution

More information

Multiobjective Evolutionary Algorithms. Pareto Rankings

Multiobjective Evolutionary Algorithms. Pareto Rankings Monografías del Semin. Matem. García de Galdeano. 7: 7 3, (3). Multiobjective Evolutionary Algorithms. Pareto Rankings Alberto, I.; Azcarate, C.; Mallor, F. & Mateo, P.M. Abstract In this work we present

More information

Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators

Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators Andrea Mambrini 1 University of Birmingham, Birmingham UK 6th June 2013 1 / 33 Andrea Mambrini GSGP: theory-laden

More information

[read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] General-to-specific ordering over hypotheses

[read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] General-to-specific ordering over hypotheses 1 CONCEPT LEARNING AND THE GENERAL-TO-SPECIFIC ORDERING [read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] Learning from examples General-to-specific ordering over hypotheses Version spaces and

More information

Automatic Differentiation Equipped Variable Elimination for Sensitivity Analysis on Probabilistic Inference Queries

Automatic Differentiation Equipped Variable Elimination for Sensitivity Analysis on Probabilistic Inference Queries Automatic Differentiation Equipped Variable Elimination for Sensitivity Analysis on Probabilistic Inference Queries Anonymous Author(s) Affiliation Address email Abstract 1 2 3 4 5 6 7 8 9 10 11 12 Probabilistic

More information

Machine Learning 2010

Machine Learning 2010 Machine Learning 2010 Concept Learning: The Logical Approach Michael M Richter Email: mrichter@ucalgary.ca 1 - Part 1 Basic Concepts and Representation Languages 2 - Why Concept Learning? Concepts describe

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

Introduction to machine learning. Concept learning. Design of a learning system. Designing a learning system

Introduction to machine learning. Concept learning. Design of a learning system. Designing a learning system Introduction to machine learning Concept learning Maria Simi, 2011/2012 Machine Learning, Tom Mitchell Mc Graw-Hill International Editions, 1997 (Cap 1, 2). Introduction to machine learning When appropriate

More information

URBAN LAND COVER AND LAND USE CLASSIFICATION USING HIGH SPATIAL RESOLUTION IMAGES AND SPATIAL METRICS

URBAN LAND COVER AND LAND USE CLASSIFICATION USING HIGH SPATIAL RESOLUTION IMAGES AND SPATIAL METRICS URBAN LAND COVER AND LAND USE CLASSIFICATION USING HIGH SPATIAL RESOLUTION IMAGES AND SPATIAL METRICS Ivan Lizarazo Universidad Distrital, Department of Cadastral Engineering, Bogota, Colombia; ilizarazo@udistrital.edu.co

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Fall 2015 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

More information

Structure Design of Neural Networks Using Genetic Algorithms

Structure Design of Neural Networks Using Genetic Algorithms Structure Design of Neural Networks Using Genetic Algorithms Satoshi Mizuta Takashi Sato Demelo Lao Masami Ikeda Toshio Shimizu Department of Electronic and Information System Engineering, Faculty of Science

More information

Inference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples

Inference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples Published in: Journal of Global Optimization, 5, pp. 69-9, 199. Inference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples Evangelos Triantaphyllou Assistant

More information

A Genetic Algorithm Approach for Doing Misuse Detection in Audit Trail Files

A Genetic Algorithm Approach for Doing Misuse Detection in Audit Trail Files A Genetic Algorithm Approach for Doing Misuse Detection in Audit Trail Files Pedro A. Diaz-Gomez and Dean F. Hougen Robotics, Evolution, Adaptation, and Learning Laboratory (REAL Lab) School of Computer

More information

Machine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20.

Machine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20. 10-601 Machine Learning, Midterm Exam: Spring 2008 Please put your name on this cover sheet If you need more room to work out your answer to a question, use the back of the page and clearly mark on the

More information

Individual learning and population evolution

Individual learning and population evolution Genetic Algorithms [Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4] Evolutionary computation Prototypical GA An example: GABIL Genetic Programming Individual learning and population evolution 168 lecture

More information

Bayesian Reasoning and Recognition

Bayesian Reasoning and Recognition Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIAG 2 / osig 1 Second Semester 2013/2014 Lesson 12 28 arch 2014 Bayesian Reasoning and Recognition Notation...2 Pattern Recognition...3

More information

Relation between Pareto-Optimal Fuzzy Rules and Pareto-Optimal Fuzzy Rule Sets

Relation between Pareto-Optimal Fuzzy Rules and Pareto-Optimal Fuzzy Rule Sets Relation between Pareto-Optimal Fuzzy Rules and Pareto-Optimal Fuzzy Rule Sets Hisao Ishibuchi, Isao Kuwajima, and Yusuke Nojima Department of Computer Science and Intelligent Systems, Osaka Prefecture

More information

Advanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland

Advanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland EnviroInfo 2004 (Geneva) Sh@ring EnviroInfo 2004 Advanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland Mikhail Kanevski 1, Michel Maignan 1

More information

Machine Learning 2010

Machine Learning 2010 Machine Learning 2010 Decision Trees Email: mrichter@ucalgary.ca -- 1 - Part 1 General -- 2 - Representation with Decision Trees (1) Examples are attribute-value vectors Representation of concepts by labeled

More information

Evolutionary Functional Link Interval Type-2 Fuzzy Neural System for Exchange Rate Prediction

Evolutionary Functional Link Interval Type-2 Fuzzy Neural System for Exchange Rate Prediction Evolutionary Functional Link Interval Type-2 Fuzzy Neural System for Exchange Rate Prediction 3. Introduction Currency exchange rate is an important element in international finance. It is one of the chaotic,

More information

Application of a GA/Bayesian Filter-Wrapper Feature Selection Method to Classification of Clinical Depression from Speech Data

Application of a GA/Bayesian Filter-Wrapper Feature Selection Method to Classification of Clinical Depression from Speech Data Application of a GA/Bayesian Filter-Wrapper Feature Selection Method to Classification of Clinical Depression from Speech Data Juan Torres 1, Ashraf Saad 2, Elliot Moore 1 1 School of Electrical and Computer

More information

Performance Evaluation and Comparison

Performance Evaluation and Comparison Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation

More information

Data Mining and Machine Learning (Machine Learning: Symbolische Ansätze)

Data Mining and Machine Learning (Machine Learning: Symbolische Ansätze) Data Mining and Machine Learning (Machine Learning: Symbolische Ansätze) Learning Individual Rules and Subgroup Discovery Introduction Batch Learning Terminology Coverage Spaces Descriptive vs. Predictive

More information

An Improved Quantum Evolutionary Algorithm with 2-Crossovers

An Improved Quantum Evolutionary Algorithm with 2-Crossovers An Improved Quantum Evolutionary Algorithm with 2-Crossovers Zhihui Xing 1, Haibin Duan 1,2, and Chunfang Xu 1 1 School of Automation Science and Electrical Engineering, Beihang University, Beijing, 100191,

More information

Metaheuristics and Local Search. Discrete optimization problems. Solution approaches

Metaheuristics and Local Search. Discrete optimization problems. Solution approaches Discrete Mathematics for Bioinformatics WS 07/08, G. W. Klau, 31. Januar 2008, 11:55 1 Metaheuristics and Local Search Discrete optimization problems Variables x 1,...,x n. Variable domains D 1,...,D n,

More information

Concept Learning. Space of Versions of Concepts Learned

Concept Learning. Space of Versions of Concepts Learned Concept Learning Space of Versions of Concepts Learned 1 A Concept Learning Task Target concept: Days on which Aldo enjoys his favorite water sport Example Sky AirTemp Humidity Wind Water Forecast EnjoySport

More information

Pattern Recognition System with Top-Down Process of Mental Rotation

Pattern Recognition System with Top-Down Process of Mental Rotation Pattern Recognition System with Top-Down Process of Mental Rotation Shunji Satoh 1, Hirotomo Aso 1, Shogo Miyake 2, and Jousuke Kuroiwa 3 1 Department of Electrical Communications, Tohoku University Aoba-yama05,

More information

Final Exam, Fall 2002

Final Exam, Fall 2002 15-781 Final Exam, Fall 22 1. Write your name and your andrew email address below. Name: Andrew ID: 2. There should be 17 pages in this exam (excluding this cover sheet). 3. If you need more room to work

More information

Lecture Notes in Machine Learning Chapter 4: Version space learning

Lecture Notes in Machine Learning Chapter 4: Version space learning Lecture Notes in Machine Learning Chapter 4: Version space learning Zdravko Markov February 17, 2004 Let us consider an example. We shall use an attribute-value language for both the examples and the hypotheses

More information

A Genetic Algorithm with Expansion and Exploration Operators for the Maximum Satisfiability Problem

A Genetic Algorithm with Expansion and Exploration Operators for the Maximum Satisfiability Problem Applied Mathematical Sciences, Vol. 7, 2013, no. 24, 1183-1190 HIKARI Ltd, www.m-hikari.com A Genetic Algorithm with Expansion and Exploration Operators for the Maximum Satisfiability Problem Anna Gorbenko

More information

An Approach to Classification Based on Fuzzy Association Rules

An Approach to Classification Based on Fuzzy Association Rules An Approach to Classification Based on Fuzzy Association Rules Zuoliang Chen, Guoqing Chen School of Economics and Management, Tsinghua University, Beijing 100084, P. R. China Abstract Classification based

More information

MODULE -4 BAYEIAN LEARNING

MODULE -4 BAYEIAN LEARNING MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities

More information

Ensemble determination using the TOPSIS decision support system in multi-objective evolutionary neural network classifiers

Ensemble determination using the TOPSIS decision support system in multi-objective evolutionary neural network classifiers Ensemble determination using the TOPSIS decision support system in multi-obective evolutionary neural network classifiers M. Cruz-Ramírez, J.C. Fernández, J. Sánchez-Monedero, F. Fernández-Navarro, C.

More information

Evaluation. Andrea Passerini Machine Learning. Evaluation

Evaluation. Andrea Passerini Machine Learning. Evaluation Andrea Passerini passerini@disi.unitn.it Machine Learning Basic concepts requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain

More information

Forecasting & Futurism

Forecasting & Futurism Article from: Forecasting & Futurism December 2013 Issue 8 A NEAT Approach to Neural Network Structure By Jeff Heaton Jeff Heaton Neural networks are a mainstay of artificial intelligence. These machine-learning

More information

Selection of Classifiers based on Multiple Classifier Behaviour

Selection of Classifiers based on Multiple Classifier Behaviour Selection of Classifiers based on Multiple Classifier Behaviour Giorgio Giacinto, Fabio Roli, and Giorgio Fumera Dept. of Electrical and Electronic Eng. - University of Cagliari Piazza d Armi, 09123 Cagliari,

More information

Computational Learning Theory. Definitions

Computational Learning Theory. Definitions Computational Learning Theory Computational learning theory is interested in theoretical analyses of the following issues. What is needed to learn effectively? Sample complexity. How many examples? Computational

More information

On Improving the k-means Algorithm to Classify Unclassified Patterns

On Improving the k-means Algorithm to Classify Unclassified Patterns On Improving the k-means Algorithm to Classify Unclassified Patterns Mohamed M. Rizk 1, Safar Mohamed Safar Alghamdi 2 1 Mathematics & Statistics Department, Faculty of Science, Taif University, Taif,

More information

Making Our Cities Safer: A Study In Neighbhorhood Crime Patterns

Making Our Cities Safer: A Study In Neighbhorhood Crime Patterns Making Our Cities Safer: A Study In Neighbhorhood Crime Patterns Aly Kane alykane@stanford.edu Ariel Sagalovsky asagalov@stanford.edu Abstract Equipped with an understanding of the factors that influence

More information

Evaluation requires to define performance measures to be optimized

Evaluation requires to define performance measures to be optimized Evaluation Basic concepts Evaluation requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain (generalization error) approximation

More information

A Genetic Inspired Optimization for ECOC

A Genetic Inspired Optimization for ECOC A Genetic Inspired Optimization for ECOC Miguel Ángel Bautista1,2, Sergio Escalera 1,2, Xavier Baró 2,3, and Oriol Pujol 1,2 1 Dept. Matemàtica Aplicada i Anàlisi, Universitat de Barcelona, Gran Via 585,

More information

Improved TBL algorithm for learning context-free grammar

Improved TBL algorithm for learning context-free grammar Proceedings of the International Multiconference on ISSN 1896-7094 Computer Science and Information Technology, pp. 267 274 2007 PIPS Improved TBL algorithm for learning context-free grammar Marcin Jaworski

More information

Condition Monitoring of Single Point Cutting Tool through Vibration Signals using Decision Tree Algorithm

Condition Monitoring of Single Point Cutting Tool through Vibration Signals using Decision Tree Algorithm Columbia International Publishing Journal of Vibration Analysis, Measurement, and Control doi:10.7726/jvamc.2015.1003 Research Paper Condition Monitoring of Single Point Cutting Tool through Vibration

More information

Text Categorization CSE 454. (Based on slides by Dan Weld, Tom Mitchell, and others)

Text Categorization CSE 454. (Based on slides by Dan Weld, Tom Mitchell, and others) Text Categorization CSE 454 (Based on slides by Dan Weld, Tom Mitchell, and others) 1 Given: Categorization A description of an instance, x X, where X is the instance language or instance space. A fixed

More information

Predictive analysis on Multivariate, Time Series datasets using Shapelets

Predictive analysis on Multivariate, Time Series datasets using Shapelets 1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,

More information

Data Mining and Machine Learning

Data Mining and Machine Learning Data Mining and Machine Learning Concept Learning and Version Spaces Introduction Concept Learning Generality Relations Refinement Operators Structured Hypothesis Spaces Simple algorithms Find-S Find-G

More information

Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network

Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network Marko Tscherepanow and Franz Kummert Applied Computer Science, Faculty of Technology, Bielefeld

More information

Review of Lecture 1. Across records. Within records. Classification, Clustering, Outlier detection. Associations

Review of Lecture 1. Across records. Within records. Classification, Clustering, Outlier detection. Associations Review of Lecture 1 This course is about finding novel actionable patterns in data. We can divide data mining algorithms (and the patterns they find) into five groups Across records Classification, Clustering,

More information

Stochastic Search: Part 2. Genetic Algorithms. Vincent A. Cicirello. Robotics Institute. Carnegie Mellon University

Stochastic Search: Part 2. Genetic Algorithms. Vincent A. Cicirello. Robotics Institute. Carnegie Mellon University Stochastic Search: Part 2 Genetic Algorithms Vincent A. Cicirello Robotics Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 cicirello@ri.cmu.edu 1 The Genetic Algorithm (GA)

More information

NEUROEVOLUTION. Contents. Evolutionary Computation. Neuroevolution. Types of neuro-evolution algorithms

NEUROEVOLUTION. Contents. Evolutionary Computation. Neuroevolution. Types of neuro-evolution algorithms Contents Evolutionary Computation overview NEUROEVOLUTION Presenter: Vlad Chiriacescu Neuroevolution overview Issues in standard Evolutionary Computation NEAT method Complexification in competitive coevolution

More information

Evolutionary Computation: introduction

Evolutionary Computation: introduction Evolutionary Computation: introduction Dirk Thierens Universiteit Utrecht The Netherlands Dirk Thierens (Universiteit Utrecht) EC Introduction 1 / 42 What? Evolutionary Computation Evolutionary Computation

More information

Machine Learning, Fall 2011: Homework 5

Machine Learning, Fall 2011: Homework 5 0-60 Machine Learning, Fall 0: Homework 5 Machine Learning Department Carnegie Mellon University Due:??? Instructions There are 3 questions on this assignment. Please submit your completed homework to

More information

COMP9444: Neural Networks. Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization

COMP9444: Neural Networks. Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization : Neural Networks Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization 11s2 VC-dimension and PAC-learning 1 How good a classifier does a learner produce? Training error is the precentage

More information

Evolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia)

Evolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia) Evolutionary Computation DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia) andrea.roli@unibo.it Evolutionary Computation Inspiring principle: theory of natural selection Species face

More information

Examination Artificial Intelligence Module Intelligent Interaction Design December 2014

Examination Artificial Intelligence Module Intelligent Interaction Design December 2014 Examination Artificial Intelligence Module Intelligent Interaction Design December 2014 Introduction This exam is closed book, you may only use a simple calculator (addition, substraction, multiplication

More information

CS 380: ARTIFICIAL INTELLIGENCE MACHINE LEARNING. Santiago Ontañón

CS 380: ARTIFICIAL INTELLIGENCE MACHINE LEARNING. Santiago Ontañón CS 380: ARTIFICIAL INTELLIGENCE MACHINE LEARNING Santiago Ontañón so367@drexel.edu Summary so far: Rational Agents Problem Solving Systematic Search: Uninformed Informed Local Search Adversarial Search

More information

CptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1

CptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1 CptS 570 Machine Learning School of EECS Washington State University CptS 570 - Machine Learning 1 IEEE Expert, October 1996 CptS 570 - Machine Learning 2 Given sample S from all possible examples D Learner

More information

Searching the Subsumption Lattice by a Genetic Algorithm

Searching the Subsumption Lattice by a Genetic Algorithm Searching the Subsumption Lattice by a Genetic Algorithm Alireza Tamaddoni-Nezhad and Stephen H. Muggleton Department of Computer Science University of York, York, YO1 5DD, UK {alireza,stephen}@cs.york.ac.uk

More information

PAC Learning Introduction to Machine Learning. Matt Gormley Lecture 14 March 5, 2018

PAC Learning Introduction to Machine Learning. Matt Gormley Lecture 14 March 5, 2018 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University PAC Learning Matt Gormley Lecture 14 March 5, 2018 1 ML Big Picture Learning Paradigms:

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

A Note on Crossover with Interval Representations

A Note on Crossover with Interval Representations A Note on Crossover with Interval Representations Christopher Stone and Larry Bull UWE Learning Classifier System Technical Report UWELCSG03-00 Faculty of Computing, Engineering and Mathematical Sciences

More information

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier Seiichi Ozawa, Shaoning Pang, and Nikola Kasabov Graduate School of Science and Technology, Kobe

More information

A new Approach to Drawing Conclusions from Data A Rough Set Perspective

A new Approach to Drawing Conclusions from Data A Rough Set Perspective Motto: Let the data speak for themselves R.A. Fisher A new Approach to Drawing Conclusions from Data A Rough et Perspective Zdzisław Pawlak Institute for Theoretical and Applied Informatics Polish Academy

More information

Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups

Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups Contemporary Mathematics Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups Robert M. Haralick, Alex D. Miasnikov, and Alexei G. Myasnikov Abstract. We review some basic methodologies

More information

Data Mining und Maschinelles Lernen

Data Mining und Maschinelles Lernen Data Mining und Maschinelles Lernen Ensemble Methods Bias-Variance Trade-off Basic Idea of Ensembles Bagging Basic Algorithm Bagging with Costs Randomization Random Forests Boosting Stacking Error-Correcting

More information

CS 231A Section 1: Linear Algebra & Probability Review

CS 231A Section 1: Linear Algebra & Probability Review CS 231A Section 1: Linear Algebra & Probability Review 1 Topics Support Vector Machines Boosting Viola-Jones face detector Linear Algebra Review Notation Operations & Properties Matrix Calculus Probability

More information