THE PERFORMANCE OF CLASSIFICATION RULES PRODUCED BY SYMBOLIC INDUCTIVE METHODS
|
|
- Shannon Byrd
- 6 years ago
- Views:
Transcription
1 Six:h inlt!rnajiona/ Symposium on.~fe/hod%gies jor ifllelligefict! SYSlt!ms Char/oue,.vC.. OClober 16 19, 1991 ~/-t5 '7/-~ USI~G GE~ETIC ALGORITHi\1S TO IMPROVE THE PERFORMANCE OF CLASSIFICATION RULES PRODUCED BY SYMBOLIC INDUCTIVE METHODS Jerzy Bala Kenneth Dejong Peter Pachowicz Center for Artificial Intelligence, George Mason University 4400 University Drive, Fairfax, VA Abstract: In this paper we present a novel way 0/ combining symbolic inductive methods and genetic algorithms (GAs) applied to produce high-performance classification rules. The presented method consists 0/ two phases. In the first one the algorithm induvtively learns a set of classification rules from noisy input examples. In the second phase the worst performing rule is optimized by GAs techniques. Experimental results are presented/or twelve classes o/noisy data obtained/rom textured images. 1 Introduction One fundamental weakness of inductive learning (except for special cases) is the fact that the acquired knowledge cannot be validated. Traditional inquiries into inductive inference have therefore dealt with questions of what are the best criteria for guiding the selection of inductive assertions, and how these assertiops can be confirmed. The goal of inference is to formulate plausible general assertions that explain the given facts and are able to predict new facts. For a given set of facts, a potentially infinite number of hypotheses can be generated that imply these facts. A preference criterion is therefore necessary to provide constraints and reduce the infinite choice to one hypothesis or a few more preferable ones. A typical way of defining such criterion is to specify the preferable propenies of the hypothesis, for example, to require that the hypothesis is the shonest or the most economical description consistent with all the facts. Even if the preferred criterion is defined it can generate other problems: e.g. performance on the future data. This is especially imponant if the initial data has some distribution of noisy and irrelevant examples. We propose a novel hybrid approach in which, frrst, a classic rule induction system (AQ) produces shon, simple, complete and, consistent rules; and then GAs are used to improve the performance of the rule via an evolutionary mechanism. By dropping the requirement for a rule
2 to be complete and consistent with respect to the initial learning dara, we reduce the coverage of noisy and irrelevant components in the rule structure. Since the AQ learning algorithm is the induction learning method in our method the next section presents a short overview of this learning method. Section 3 presents the GA phase of our method and the experimental results are presented in section 4. 2 AQ learning algorithm The AQ algorithm [Michalski et al., 1986] learns atuibutional description from examples. When building a decision rule. AQ performs a heuristic search through a space of logical expressions to determine those that account for all positive examples and no negative examples. Because there are usually many such complete and consistent expressions, the goal of AQ is to find the most preferred one, according to some criterion. Learning examples are given in the form of events. which are vectors of attribute values. Events represent different decision classes. For each class a decision rule is produced that covers all examples of that class and none of other classes. The concept descriptions learned by AQ algorithm are represented in VL 1, the Variable Valued Logic System 1 [Michalski. 1972]. and are used to represent attributional concept descriptions. A description of a concept is a disjunctive normal form which is called a cover. A cover is a disjunction of complexes. A complex is a conjunction of selectors. A selector is a form: [L#R] (1) where. L is called the referee. which is an attribute. R is called the referent, which is a set of values in the domain of the attribute L. # is one of the following relational symbols: =, <, >. >=, <=, <>. The following is an example of the five selectors AQ complex (equality is used as a relational symbol): [xl=1.. 3][x2=1][x4=O][x6=1..7][x8=1] Dots in the description represent a range of possible values of a given attribute. An example of simple, four complexes. AQ description is presented below: 1 2 [xl=7..8] [x2=8.. 19] [x ] [x5=4..54] [x6=o..3] [xl= ] [x3=11..14] [x4= ] [x6=o.. 9] [x7=o [xl=9.. 18] [x3=16..21] [x4=9.. 10] [xl=1o.. 14] [x3= ] [x4= ] [x7=4.. 5] 2
3 3 Genetic Algorithm's Phase The rules generated by the AQ algorithm have the propeny of being complete and consistent with respect to the learning examples. These properties mean that each class covers its learning examples (positive examples) and does not cover any negative examples (negative examples are all examples belonging to other classes), \\'hen the AQ algorithm is applied to finergrained and noisy symbolic representations the consistency and completeness conditions overconstrain the generated class description. This, in turn, leads to poorer predictive accuracy on future data. Conversely, by relaxing the requirement of consistency and completeness, predictive accuracy in noisy domains can be improved. In the proposed method we use GAs to improve the performance of initially consistent and complete rules via the evolutionary mechanism. A genetic algorithm [De Jong. 1988] maintains a constant-sized population of candidate solutions. known as individuals. At each iteration, each individual is evaluated and recombined with others on the basis of its overall quality or fitness. The expected number of times an individual is selected for recombination is proportional to its fitness relative to the rest of population. The power of a genetic algorithm lies in its ability to exploit, in a highly efficient manner, information about a large number of individuals. By allocating more reproductive occurrences to above average individuals, the overall affect is an increase of the population's average fitness. New individuals are created using two main genetic recombination operators known as crossover and mutation. Crossover operates by selecting a random location in the genetic saing of the parents (crossover point) and concatenating the initial segments of one parent with the final segment parent to create a new child. A second child is simultaneously generated using the remaining segments of the two parents. Mutation provides for occasional disturbances in the crossover operation to insure diversity in the genetic strings over long periods of rime and to prevent stagnation in the convergence of the optimization's technique. In order to apply GAs. one needs to choose a representation, defme the operators and the performance measure. In traditional GA's individuals in the population are typically represented using a binary string notation to promote efficiency and application independence in the genetic operations. The mathematical analysis of GAs shows that they work best when the internal representation encourages the emergence of useful building blocks that can be subsequently combined with others to produce improved performance and string representations are just one of many ways of achieving this. In old' application of GAs we do not use string representations. Each individual of the population is represented in VLl (previous section). as the different version of a given rule that is a subject for the optimization. The VL 1 representation is very "natural" for the optimization problem defined in this paper (search for a better performing rule) and can easily be manipulated by GA operators. The next section describes GA operators chosen for our method. 3
4 3.1 GA operators An initial population of rules is created by causing small variations to an existing rule. Then, a population of different variations of this rule is evolved by using GA's operators. Each rule in the population is evaluated based on its peiiormance within the set of initial rules. To evaluate the rule we use the tuning data which are pan of the set of learning examples. The perfonnance of the best rule in the population is monitored by matching this rule with the testing examples. Mutation is performed by introducing small changes to the condition pans (selectors) of a selected disjunct (rule complex) of a given rule. Th,e selectors to be changed are chosen by randomly generating two pointers. the first one that points to the rule disjunct complex. and the second one points to the selector of this disjunct. Random changes are introduced to the left-most or right-most (this is also chosen randomly) value of this selector. For example, the selector [x5= ] can be changed to one of the following; [x5=3,io [x5=3,io..241, [x5= ] or [x5= ]. Such a mutation process samples the space of possible boundaries between rules to minimize the coverage of noisy and irrelevant components of a given rule. The crossover operation is performed by splitting rule disjuncts into two parts, upper disjuncts and lower disjuncts. These parts are exchanged between parent rules to produce new child rules. Since degree of match of a given instance depends on the degree of match of this instance to each disjunct of that rule, this exchange process enables inheritance of information about strong disjuncts (suongly matching) in the individuals of the next evolved population. An example ofcrossover applied to short, four disjuncts rules is depicted below: Parent rule I I [xl=7..8] [x2=8.. 19] [x3=8.. 13] [xs=4..54] [x6=o..3] 2 [xl=is.. 54] [x3=11..14] [x4= ] [x6=o.. 9] [x7=o..11] crossover position [xl=9..18] [x3=16..21] [x4=9..10] 4 [xl=io.. 14] [x3= ] [x4=14..54] [x7=4.. 5] Parent rule 2 I [x3= ] [x4= ] [xs=o.. 6] [x7=5.. 12] 2 [xl=8.. 25] [x3=8.. 13] [x4=9.. 11] [x5=o.. 3] crossover position [x4=o. 22] [x5=8.. 9] [x6=o.. 7] [x7=11..48] 4 [x2=s. 8] [x3=7.. 8] [x4=8.. 11] [x5=o.. 3] Result of the crossover operation (one of two child rules) I [xl=7. 8] [x2=8.. 19] [x3=8.. 13] [xs=4..54] [x6=o..3] 2 [xl=is.. 54] [x3=11..14] [x4= ] [x6=o..9] [x7=o.. 11] 3 [x4=o.. 22] [x5=8.. 9] [x6=o.. 7] [x7=11..48] 4 [x2=5.. 8] [x3=7.. 8] [x4=8.. 11] [xs=o. 3] 4
5 The GA operators as described above are very effective in situations where no a priori infonnation is given on the distribution of attribute values. Son-nonnal distribution of attribute values complicates the AQ generated descriptions, especially if the learned rules consistently and completely cover input instances of the data. Graph 1 presents the most representative samples of non-normal attribute distribution obtained from the experiment with textured image classes described in section 4. Class: C9; Attribute:,,5 8~--~----,---~----,---~~ Class: C8; Attribute: xl O;--'~~~-r----r-~-+~~~ O~~--~---+--P-~~--+--r-;~ o 10 AUribute value o 10 AUribute value Class: C8; Attribute: x3 ~.g -l t-_,r--ii--1t a ~ -+--I-----t'--+--tfo~ + t i+---~--~~~--+---~~--~~ l+---~~--r---~--~~r-~ Class: Cll; Attribute:...7 4,-----~--~----~----~--~~ O~~--+--p--P-~~ ~~~ o 10 AUribuae value O~~~--~~~~--~~-r~~ o 10 AUribute value Graph 1. Examples of non-nonnal attribute distribution. The solid line corresponds to smoothed distribution of an attribute and the dotted line corresponds to the approximated nonnal distribution. The left hand diagrams of Graph 1 present a relatively simple attribute distribution. However, the right hand diagrams of Graph I present two cases of more complex attribute distributions. The upper right diagram indicates the possible formation of more than one cluster of training data. The lower right diagram presents very complex distribution of an attribute without distinction of regular clusters. One can notice that x7 5
6 attribute does not carry significant information about CII class. Considering the above complexity of the attributes. we postulate that this domain (data in a representation space expressed by such attributes) is a good candidate for the GA type of searches, where random changes (mutation) and some directional sampling (crossover) can yield high performance solutions (optimized rules). 3.2 Rule performance evaluation Each rule candidate in the population is evaluated by plugging it into the set of other rules and calculating the confusion mattix for the tuning examples. The confusion matrix represents infonnation on correct and incorrect classification results. This matrix is obtained by calculating a degree of flexible match between a tuning instance and a given class description. The row entries it) the matrix represent a percentage of matched instances from a given class (row index) to all class descriptions (column index). The following is an example of the confusion matrix for 12 classes: Cl C2 C3 C4 C5 C6 Cl C8 C9 CIO Cl1 C12 Cl 45 S C2 S C S C C C C C C I 2 CI0 0 0 I Cll C Average recognition=74.67% Average mis-classitication=4.83 The above confusion matrix represents classification results on data extracted from textured images. The method of extraction and texture classification by inductive learning is not the subject of research presented in this paper. The reader can find various information on the method of learning texture descriptions in the following papers [Bala and Pachowicz. 1991]. Let us suppose that the class Cll is the one chosen to be evolved by GAs mechanism. First we mutate this class representation multiple times to produce an initial population of different individuals of that class. To test the performance of a given individual we calculate the confusion matrix. The confusion matrix shows us how this specific variation of class Cll 6
7 perfonns when tested with other classes using the set of tuning examples. For example. we can see from me above confusion matrix mat the class C 11 has me following performance: Cl C2 C3 C4 C5 C6 C7 C8 C9 ClO Cll Cl2 CII The above values represent percentage of instances of class C 11 (set of tuning instances) matched to learned descriptions of all 12 classes. Ifclass CII is to be optimized. by GAs we have to evaluate the performance of each individual from the population of that class by calculating the performance measure of different candidates of that class with respect to other classes. Thus, as the performance evaluation measure of each individual of the population we use the ratio of correct recognition rates to incorrect recognition rates; CC/MC, where CC is an average recognition for correctly recognized classes (an average of entries on the diagonal of the confusion matrix). and MC is an average mis-classification (an average of entries outside the diagonal of the confusion matrix). For the above confusion matrix CC/MC=74.67/4.83= The next section describes how the distance measure between an example (tuning or testing) and a rule is calculated Reeo&nizin& class membership throu&h nexible matehin& Flexible matching measures the degree of closeness between an instance and the conditional pan of a rule. Such closeness is computed as distance value between an instance and rule complex within the attribute space. This distance measure takes any value in the range from o (Le., does not match) to 1 (Le. matches). Calculation of this distance measure for a single test instance is executed. according to the following schema. For a given condition ofa rule: [ Xn =valjl and an instance where Xn = valk ' the normalized distance measure is 1 ( I val j val k II #levels ) (2) where #levels is the total number of attrib\1te values. The confidence level of a rule is computed by multiplying evaluation values of each condition of the rule. The total evaluation of class membership of a given test instance is equal to the confidence level of the best matching rule, i.e. the rule of the highest confidence value. For example. the confidence value c for matching the following complex of a rule [ xl =0 ] [ x7 =10 ] [ x8 = ] with a given instance x = < 4. S. 24, 34, O. 12, 6, 25 > #levels=55 is computed as follows: and the number of attribute values c x } = 1 ( 10-41/55) =.928 C x 7 = 1 - ( / 55) =.928 c x 8 = 1 - ( ) =.91 Cx2, cx3. cx4' cxs, cx6 = 1 7
8 c = cxl * Cx2 * Cx) * Cx4 * Cx5 * Cx6 * Cx7 * C x 8 = 0.78 The recognition process yields class membership for which the confidence level is the highest among matched rules. Calculated confidence level, however, is not a probability measure and it can yield more than one class membership. It means that for a given preclassified test dataset. the system recognition effectiveness is calculated as the number of correctly classified instances to the total number of instances of the dataset. The above described method of rule evaluation by flexible matching and confidence level calculation has imponant consequences for the crossover operation. Since the degree of match of an instance to a rule depends on the confidence level of this instance to each disjunct of that rule, the swapping process introduced by crossover operator enables inheritance of infonnation represented in the strong disjuncts (with high confidence level) of rule candidates through generations of a population. 4. Experimental results In the experiment we used 12 classes of texture data. The initial descriptions were generated by the AQ module using 200 examples per class. Another set of 100 examples was used to calculat!' the perfonnance measure for the GAs cycle. The testing set had 200 examples extracted from different image areas other than training and tuning data.. Examples of a given class were constructed using 8 attributes and an extraction process based on Law's mask [Bala and Pachowicz, 1991]. The "weakest" class (with the lowest matching results with testing data) chosen for the experiment was class Cll (like in the confusion matrix example in section 3.2) with 40 disjuncts. Graph 2 represents results of the experiment. White circles of the diagrams represent characteristics obtained for tuning data used to guide the genetic search. Characteristics mapped by black circles were obtained for testing data. Upper diagram monitors the performance of genetically evolved description of Cll texture. When all 300 examples were used to generate rules, the average classification rate for this class (when tested with 200 examples) was below 45%. When the set of 300 examples was split into two parts, 200 for the initial inductive learning and 100 as the tuning data for GAs cycle, the correct classification rates obtainedtin the 30th evolution was above 60%. That is a significant increase in comparison with 45% obtained from inductive learning only. The bottom diagrams represent the evaluation function used by genetic algorithm (COMe, section 3.2) in order to guide the genetic sean:h. The evaluation function was calculated as a rate of the correct classifications to mis-classifications for all twelve texture classes. These diagrams are depicted for both testing and tuning data. The increases of CC/MC on both diagrams represent an overall improvement of system recognition performance. The system performance was investigated for a larger number of GA generation steps. However, it appears, that the substantial increase was reached both for the CII class description and for the overall system performance in a very few generation steps (Le., in 10 steps). 8
9 o Results for tuning dat.a Results for testing data 30~----~A ~ 20;---~--~----~--~~--~--~ o 10 Generations ~ 0 "1 ~ r J roo II MM be, u ~ /\.J ~ f \-\ o 10 Generations Generations 30 Graph 2. Results of experiment with 12 classes. s. Conclusion This paper presents a novel approach to the optimization of the classification rule descriptions through GAs techniques. We recognize rule optimization as an emerging research area in the creation of autonomous intelligent systems. Two issues appear to be important for this research. First, rule optimization is absolutely necessary in such problems where the system must be protected against the accumulation of noisy components and where attributes used to describe initial data have complex. non-nonna! distributions. The method can be augmented by 9
10 optimizing input data formation (searching for the best subset of attributes) prior to rule generation [Vafaie, De Jong 1991]. Secondly, for inducti...ely generated rules (falsity preserving rules) there should be some method of validating these rules (or choosing the most preferred inductive rule) before applying them to future data. The methoo presented in this paper tries to offer some solutions to these problems. The presented experiments on applying this hybrid methoo to finer-grained symbolic representations of vision data are encouraging. Highly disjunctive descriptions obtained by inductive learning from this data were easily manipulated by chosen GA operators and a substantial increase in the overall system perfonnance was observed in a few generations. Although there exists substantial differences between GAs and symbolic induction methods (learning algorithm, performance elements, knowledge representation), this method serves as a promising example that a better understanding of abilities of each approach (identifying common building blocks) can lead to novel and useful ways of combining them. This paper concentrates primarily on the proposed methodology. More experiments are needed to draw concrete conclusions. Future research will attempt to thoroughly test this approach on numerous data sets. We will also examine more thoroughly mutation and crossover operators. Acknowledgement: This research was supponed in pan by the Office of Naval Research under grants No. NOOOI4-88-K-0397 and No. NOOO14-88-K-0226, and in pan by the Defense Advan~ed Research Projects Agency under the grant administered by the Office of Naval Research No. NOOOI4-K The authors wish to thank: Janet Holmes and Haleh Vafaie for editing suggestions. References Bala. I.W. and Pachowicz, P.W., "Application of Symbolic Machine Learning to the Recognition oftexture Concepts" 7th IEEE Conference on Artificial Intelligence Applications, Miami Beach FL, February De Jong, K "Learning with Genetic Algorithms: An Overview". Machine Learning vol. 3; , Kluwer Academic Publishers Vafaie H.and De Jong, K.,"Improving the Performance of a Rule Induction System Using Genetic Algorithms"; in preparation.(center for AI, GMU 1991) Fitzpatrick. J.M. and Grefenstette I., "Genetic Algorithms in Noisy Environments", Machine Leaming 3 pp (1988). MichalskiR. S. "AQVAUI--ComputerImplementation ofa Variable-Valued Logic System VLl and Examples of Its Application to Pattern Recognition", First International Joint Conference on Pattern Recognition, October 30, 1973, Washington D.C. Michalski, R.S. Mozetic I., Hong I.R., Lavrac N. "The AQ15 Inductive Learning System", Report No. UIUCDCS-R , Department of Computer Science, University of Illinois at Urbane-Champaign, July
Learning From Inconsistent and Noisy Data: The AQ18 Approach *
Eleventh International Symposium on Methodologies for Intelligent Systems, Warsaw, pp. 411-419, 1999 Learning From Inconsistent and Noisy Data: The AQ18 Approach * Kenneth A. Kaufman and Ryszard S. Michalski*
More informationGenetic Algorithms: Basic Principles and Applications
Genetic Algorithms: Basic Principles and Applications C. A. MURTHY MACHINE INTELLIGENCE UNIT INDIAN STATISTICAL INSTITUTE 203, B.T.ROAD KOLKATA-700108 e-mail: murthy@isical.ac.in Genetic algorithms (GAs)
More informationLooking Under the EA Hood with Price s Equation
Looking Under the EA Hood with Price s Equation Jeffrey K. Bassett 1, Mitchell A. Potter 2, and Kenneth A. De Jong 1 1 George Mason University, Fairfax, VA 22030 {jbassett, kdejong}@cs.gmu.edu 2 Naval
More informationLecture 9 Evolutionary Computation: Genetic algorithms
Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Simulation of natural evolution Genetic algorithms Case study: maintenance scheduling with genetic
More informationLecture 22. Introduction to Genetic Algorithms
Lecture 22 Introduction to Genetic Algorithms Thursday 14 November 2002 William H. Hsu, KSU http://www.kddresearch.org http://www.cis.ksu.edu/~bhsu Readings: Sections 9.1-9.4, Mitchell Chapter 1, Sections
More informationCSC 4510 Machine Learning
10: Gene(c Algorithms CSC 4510 Machine Learning Dr. Mary Angela Papalaskari Department of CompuBng Sciences Villanova University Course website: www.csc.villanova.edu/~map/4510/ Slides of this presenta(on
More informationSearch. Search is a key component of intelligent problem solving. Get closer to the goal if time is not enough
Search Search is a key component of intelligent problem solving Search can be used to Find a desired goal if time allows Get closer to the goal if time is not enough section 11 page 1 The size of the search
More informationArtificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence
Artificial Intelligence (AI) Artificial Intelligence AI is an attempt to reproduce intelligent reasoning using machines * * H. M. Cartwright, Applications of Artificial Intelligence in Chemistry, 1993,
More informationChoosing Variables with a Genetic Algorithm for Econometric models based on Neural Networks learning and adaptation.
Choosing Variables with a Genetic Algorithm for Econometric models based on Neural Networks learning and adaptation. Daniel Ramírez A., Israel Truijillo E. LINDA LAB, Computer Department, UNAM Facultad
More informationAn Evolutionary Programming Based Algorithm for HMM training
An Evolutionary Programming Based Algorithm for HMM training Ewa Figielska,Wlodzimierz Kasprzak Institute of Control and Computation Engineering, Warsaw University of Technology ul. Nowowiejska 15/19,
More informationSelected Algorithms of Machine Learning from Examples
Fundamenta Informaticae 18 (1993), 193 207 Selected Algorithms of Machine Learning from Examples Jerzy W. GRZYMALA-BUSSE Department of Computer Science, University of Kansas Lawrence, KS 66045, U. S. A.
More informationDecision Trees (Cont.)
Decision Trees (Cont.) R&N Chapter 18.2,18.3 Side example with discrete (categorical) attributes: Predicting age (3 values: less than 30, 30-45, more than 45 yrs old) from census data. Attributes (split
More informationEfficiently merging symbolic rules into integrated rules
Efficiently merging symbolic rules into integrated rules Jim Prentzas a, Ioannis Hatzilygeroudis b a Democritus University of Thrace, School of Education Sciences Department of Education Sciences in Pre-School
More informationDETECTION OF ATMOSPHERIC RELEASES
J1. MACHINE LEARNING EVOLUTION FOR THE SOURCE DETECTION OF ATMOSPHERIC RELEASES G. Cervone 1 and P. Franzese 2 1 Dept. of Geography and Geoinformation Science, George Mason University 2 Center for Earth
More informationTUNING ROUGH CONTROLLERS BY GENETIC ALGORITHMS
TUNING ROUGH CONTROLLERS BY GENETIC ALGORITHMS Teresa Chiu Department of Mathematics and Computer Science Sun Jose State University San Jose, California 95192 T.Y. Lin* Department of Electrical and Computer
More informationRepresentation and Hidden Bias II: Eliminating Defining Length Bias in Genetic Search via Shuffle Crossover
Representation and Hidden Bias II: Eliminating Defining Length Bias in Genetic Search via Shuffle Crossover Abstract The traditional crossover operator used in genetic search exhibits a position-dependent
More informationAn Evolution Strategy for the Induction of Fuzzy Finite-state Automata
Journal of Mathematics and Statistics 2 (2): 386-390, 2006 ISSN 1549-3644 Science Publications, 2006 An Evolution Strategy for the Induction of Fuzzy Finite-state Automata 1,2 Mozhiwen and 1 Wanmin 1 College
More informationA Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997
A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997 Vasant Honavar Artificial Intelligence Research Laboratory Department of Computer Science
More informationAN EVOLUTIONARY ALGORITHM TO ESTIMATE UNKNOWN HEAT FLUX IN A ONE- DIMENSIONAL INVERSE HEAT CONDUCTION PROBLEM
Proceedings of the 5 th International Conference on Inverse Problems in Engineering: Theory and Practice, Cambridge, UK, 11-15 th July 005. AN EVOLUTIONARY ALGORITHM TO ESTIMATE UNKNOWN HEAT FLUX IN A
More informationPredicting Dynamical Systems
two projects in which a genetic algorithm is used to solve such search problems-one of predicting dynamical systems, and the other of predicting the structure of proteins. Predicting Dynamical Systems
More informationData Warehousing & Data Mining
13. Meta-Algorithms for Classification Data Warehousing & Data Mining Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 13.
More informationOptimization Models for Detection of Patterns in Data
Optimization Models for Detection of Patterns in Data Igor Masich, Lev Kazakovtsev, and Alena Stupina Siberian State University of Science and Technology, Krasnoyarsk, Russia is.masich@gmail.com Abstract.
More informationA. Pelliccioni (*), R. Cotroneo (*), F. Pungì (*) (*)ISPESL-DIPIA, Via Fontana Candida 1, 00040, Monteporzio Catone (RM), Italy.
Application of Neural Net Models to classify and to forecast the observed precipitation type at the ground using the Artificial Intelligence Competition data set. A. Pelliccioni (*), R. Cotroneo (*), F.
More informationGENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS
GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS A genetic algorithm is a random search technique for global optimisation in a complex search space. It was originally inspired by an
More informationSection on Survey Research Methods JSM 2010 STATISTICAL GRAPHICS OF PEARSON RESIDUALS IN SURVEY LOGISTIC REGRESSION DIAGNOSIS
STATISTICAL GRAPHICS OF PEARSON RESIDUALS IN SURVEY LOGISTIC REGRESSION DIAGNOSIS Stanley Weng, National Agricultural Statistics Service, U.S. Department of Agriculture 3251 Old Lee Hwy, Fairfax, VA 22030,
More informationThe Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan
The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan ABSTRACT The genetic algorithm is an adaptive search method that has the ability for a smart search
More informationHypothesis Evaluation
Hypothesis Evaluation Machine Learning Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Hypothesis Evaluation Fall 1395 1 / 31 Table of contents 1 Introduction
More informationCS 6375 Machine Learning
CS 6375 Machine Learning Decision Trees Instructor: Yang Liu 1 Supervised Classifier X 1 X 2. X M Ref class label 2 1 Three variables: Attribute 1: Hair = {blond, dark} Attribute 2: Height = {tall, short}
More informationLecture 2: Foundations of Concept Learning
Lecture 2: Foundations of Concept Learning Cognitive Systems II - Machine Learning WS 2005/2006 Part I: Basic Approaches to Concept Learning Version Space, Candidate Elimination, Inductive Bias Lecture
More informationMetaheuristics and Local Search
Metaheuristics and Local Search 8000 Discrete optimization problems Variables x 1,..., x n. Variable domains D 1,..., D n, with D j Z. Constraints C 1,..., C m, with C i D 1 D n. Objective function f :
More information[Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4]
1 EVOLUTIONARY ALGORITHMS [Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4] Evolutionary computation Prototypical GA An example: GABIL Schema theorem Genetic Programming Individual learning and population
More informationModeling of a Winding Machine Using Genetic Programming
Modeling of a Winding Machine Using Genetic Programming Abo El-Abbass Hussian +, Alaa Sheta +, Mahmoud Kamel ++ Mohammed Telbaney + and Ashraf Abdelwahab + + Computers and Systems Department, Electronics
More informationBinary Particle Swarm Optimization with Crossover Operation for Discrete Optimization
Binary Particle Swarm Optimization with Crossover Operation for Discrete Optimization Deepak Singh Raipur Institute of Technology Raipur, India Vikas Singh ABV- Indian Institute of Information Technology
More informationCrossover Techniques in GAs
Crossover Techniques in GAs Debasis Samanta Indian Institute of Technology Kharagpur dsamanta@iitkgp.ac.in 16.03.2018 Debasis Samanta (IIT Kharagpur) Soft Computing Applications 16.03.2018 1 / 1 Important
More informationGenetic Algorithms. Seth Bacon. 4/25/2005 Seth Bacon 1
Genetic Algorithms Seth Bacon 4/25/2005 Seth Bacon 1 What are Genetic Algorithms Search algorithm based on selection and genetics Manipulate a population of candidate solutions to find a good solution
More informationMultiobjective Evolutionary Algorithms. Pareto Rankings
Monografías del Semin. Matem. García de Galdeano. 7: 7 3, (3). Multiobjective Evolutionary Algorithms. Pareto Rankings Alberto, I.; Azcarate, C.; Mallor, F. & Mateo, P.M. Abstract In this work we present
More informationGeometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators
Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators Andrea Mambrini 1 University of Birmingham, Birmingham UK 6th June 2013 1 / 33 Andrea Mambrini GSGP: theory-laden
More information[read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] General-to-specific ordering over hypotheses
1 CONCEPT LEARNING AND THE GENERAL-TO-SPECIFIC ORDERING [read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] Learning from examples General-to-specific ordering over hypotheses Version spaces and
More informationAutomatic Differentiation Equipped Variable Elimination for Sensitivity Analysis on Probabilistic Inference Queries
Automatic Differentiation Equipped Variable Elimination for Sensitivity Analysis on Probabilistic Inference Queries Anonymous Author(s) Affiliation Address email Abstract 1 2 3 4 5 6 7 8 9 10 11 12 Probabilistic
More informationMachine Learning 2010
Machine Learning 2010 Concept Learning: The Logical Approach Michael M Richter Email: mrichter@ucalgary.ca 1 - Part 1 Basic Concepts and Representation Languages 2 - Why Concept Learning? Concepts describe
More informationPerformance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project
Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore
More informationIntroduction to machine learning. Concept learning. Design of a learning system. Designing a learning system
Introduction to machine learning Concept learning Maria Simi, 2011/2012 Machine Learning, Tom Mitchell Mc Graw-Hill International Editions, 1997 (Cap 1, 2). Introduction to machine learning When appropriate
More informationURBAN LAND COVER AND LAND USE CLASSIFICATION USING HIGH SPATIAL RESOLUTION IMAGES AND SPATIAL METRICS
URBAN LAND COVER AND LAND USE CLASSIFICATION USING HIGH SPATIAL RESOLUTION IMAGES AND SPATIAL METRICS Ivan Lizarazo Universidad Distrital, Department of Cadastral Engineering, Bogota, Colombia; ilizarazo@udistrital.edu.co
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Fall 2015 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
More informationStructure Design of Neural Networks Using Genetic Algorithms
Structure Design of Neural Networks Using Genetic Algorithms Satoshi Mizuta Takashi Sato Demelo Lao Masami Ikeda Toshio Shimizu Department of Electronic and Information System Engineering, Faculty of Science
More informationInference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples
Published in: Journal of Global Optimization, 5, pp. 69-9, 199. Inference of A Minimum Size Boolean Function by Using A New Efficient Branch-and-Bound Approach From Examples Evangelos Triantaphyllou Assistant
More informationA Genetic Algorithm Approach for Doing Misuse Detection in Audit Trail Files
A Genetic Algorithm Approach for Doing Misuse Detection in Audit Trail Files Pedro A. Diaz-Gomez and Dean F. Hougen Robotics, Evolution, Adaptation, and Learning Laboratory (REAL Lab) School of Computer
More informationMachine Learning, Midterm Exam: Spring 2008 SOLUTIONS. Q Topic Max. Score Score. 1 Short answer questions 20.
10-601 Machine Learning, Midterm Exam: Spring 2008 Please put your name on this cover sheet If you need more room to work out your answer to a question, use the back of the page and clearly mark on the
More informationIndividual learning and population evolution
Genetic Algorithms [Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4] Evolutionary computation Prototypical GA An example: GABIL Genetic Programming Individual learning and population evolution 168 lecture
More informationBayesian Reasoning and Recognition
Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIAG 2 / osig 1 Second Semester 2013/2014 Lesson 12 28 arch 2014 Bayesian Reasoning and Recognition Notation...2 Pattern Recognition...3
More informationRelation between Pareto-Optimal Fuzzy Rules and Pareto-Optimal Fuzzy Rule Sets
Relation between Pareto-Optimal Fuzzy Rules and Pareto-Optimal Fuzzy Rule Sets Hisao Ishibuchi, Isao Kuwajima, and Yusuke Nojima Department of Computer Science and Intelligent Systems, Osaka Prefecture
More informationAdvanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland
EnviroInfo 2004 (Geneva) Sh@ring EnviroInfo 2004 Advanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland Mikhail Kanevski 1, Michel Maignan 1
More informationMachine Learning 2010
Machine Learning 2010 Decision Trees Email: mrichter@ucalgary.ca -- 1 - Part 1 General -- 2 - Representation with Decision Trees (1) Examples are attribute-value vectors Representation of concepts by labeled
More informationEvolutionary Functional Link Interval Type-2 Fuzzy Neural System for Exchange Rate Prediction
Evolutionary Functional Link Interval Type-2 Fuzzy Neural System for Exchange Rate Prediction 3. Introduction Currency exchange rate is an important element in international finance. It is one of the chaotic,
More informationApplication of a GA/Bayesian Filter-Wrapper Feature Selection Method to Classification of Clinical Depression from Speech Data
Application of a GA/Bayesian Filter-Wrapper Feature Selection Method to Classification of Clinical Depression from Speech Data Juan Torres 1, Ashraf Saad 2, Elliot Moore 1 1 School of Electrical and Computer
More informationPerformance Evaluation and Comparison
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation
More informationData Mining and Machine Learning (Machine Learning: Symbolische Ansätze)
Data Mining and Machine Learning (Machine Learning: Symbolische Ansätze) Learning Individual Rules and Subgroup Discovery Introduction Batch Learning Terminology Coverage Spaces Descriptive vs. Predictive
More informationAn Improved Quantum Evolutionary Algorithm with 2-Crossovers
An Improved Quantum Evolutionary Algorithm with 2-Crossovers Zhihui Xing 1, Haibin Duan 1,2, and Chunfang Xu 1 1 School of Automation Science and Electrical Engineering, Beihang University, Beijing, 100191,
More informationMetaheuristics and Local Search. Discrete optimization problems. Solution approaches
Discrete Mathematics for Bioinformatics WS 07/08, G. W. Klau, 31. Januar 2008, 11:55 1 Metaheuristics and Local Search Discrete optimization problems Variables x 1,...,x n. Variable domains D 1,...,D n,
More informationConcept Learning. Space of Versions of Concepts Learned
Concept Learning Space of Versions of Concepts Learned 1 A Concept Learning Task Target concept: Days on which Aldo enjoys his favorite water sport Example Sky AirTemp Humidity Wind Water Forecast EnjoySport
More informationPattern Recognition System with Top-Down Process of Mental Rotation
Pattern Recognition System with Top-Down Process of Mental Rotation Shunji Satoh 1, Hirotomo Aso 1, Shogo Miyake 2, and Jousuke Kuroiwa 3 1 Department of Electrical Communications, Tohoku University Aoba-yama05,
More informationFinal Exam, Fall 2002
15-781 Final Exam, Fall 22 1. Write your name and your andrew email address below. Name: Andrew ID: 2. There should be 17 pages in this exam (excluding this cover sheet). 3. If you need more room to work
More informationLecture Notes in Machine Learning Chapter 4: Version space learning
Lecture Notes in Machine Learning Chapter 4: Version space learning Zdravko Markov February 17, 2004 Let us consider an example. We shall use an attribute-value language for both the examples and the hypotheses
More informationA Genetic Algorithm with Expansion and Exploration Operators for the Maximum Satisfiability Problem
Applied Mathematical Sciences, Vol. 7, 2013, no. 24, 1183-1190 HIKARI Ltd, www.m-hikari.com A Genetic Algorithm with Expansion and Exploration Operators for the Maximum Satisfiability Problem Anna Gorbenko
More informationAn Approach to Classification Based on Fuzzy Association Rules
An Approach to Classification Based on Fuzzy Association Rules Zuoliang Chen, Guoqing Chen School of Economics and Management, Tsinghua University, Beijing 100084, P. R. China Abstract Classification based
More informationMODULE -4 BAYEIAN LEARNING
MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities
More informationEnsemble determination using the TOPSIS decision support system in multi-objective evolutionary neural network classifiers
Ensemble determination using the TOPSIS decision support system in multi-obective evolutionary neural network classifiers M. Cruz-Ramírez, J.C. Fernández, J. Sánchez-Monedero, F. Fernández-Navarro, C.
More informationEvaluation. Andrea Passerini Machine Learning. Evaluation
Andrea Passerini passerini@disi.unitn.it Machine Learning Basic concepts requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain
More informationForecasting & Futurism
Article from: Forecasting & Futurism December 2013 Issue 8 A NEAT Approach to Neural Network Structure By Jeff Heaton Jeff Heaton Neural networks are a mainstay of artificial intelligence. These machine-learning
More informationSelection of Classifiers based on Multiple Classifier Behaviour
Selection of Classifiers based on Multiple Classifier Behaviour Giorgio Giacinto, Fabio Roli, and Giorgio Fumera Dept. of Electrical and Electronic Eng. - University of Cagliari Piazza d Armi, 09123 Cagliari,
More informationComputational Learning Theory. Definitions
Computational Learning Theory Computational learning theory is interested in theoretical analyses of the following issues. What is needed to learn effectively? Sample complexity. How many examples? Computational
More informationOn Improving the k-means Algorithm to Classify Unclassified Patterns
On Improving the k-means Algorithm to Classify Unclassified Patterns Mohamed M. Rizk 1, Safar Mohamed Safar Alghamdi 2 1 Mathematics & Statistics Department, Faculty of Science, Taif University, Taif,
More informationMaking Our Cities Safer: A Study In Neighbhorhood Crime Patterns
Making Our Cities Safer: A Study In Neighbhorhood Crime Patterns Aly Kane alykane@stanford.edu Ariel Sagalovsky asagalov@stanford.edu Abstract Equipped with an understanding of the factors that influence
More informationEvaluation requires to define performance measures to be optimized
Evaluation Basic concepts Evaluation requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain (generalization error) approximation
More informationA Genetic Inspired Optimization for ECOC
A Genetic Inspired Optimization for ECOC Miguel Ángel Bautista1,2, Sergio Escalera 1,2, Xavier Baró 2,3, and Oriol Pujol 1,2 1 Dept. Matemàtica Aplicada i Anàlisi, Universitat de Barcelona, Gran Via 585,
More informationImproved TBL algorithm for learning context-free grammar
Proceedings of the International Multiconference on ISSN 1896-7094 Computer Science and Information Technology, pp. 267 274 2007 PIPS Improved TBL algorithm for learning context-free grammar Marcin Jaworski
More informationCondition Monitoring of Single Point Cutting Tool through Vibration Signals using Decision Tree Algorithm
Columbia International Publishing Journal of Vibration Analysis, Measurement, and Control doi:10.7726/jvamc.2015.1003 Research Paper Condition Monitoring of Single Point Cutting Tool through Vibration
More informationText Categorization CSE 454. (Based on slides by Dan Weld, Tom Mitchell, and others)
Text Categorization CSE 454 (Based on slides by Dan Weld, Tom Mitchell, and others) 1 Given: Categorization A description of an instance, x X, where X is the instance language or instance space. A fixed
More informationPredictive analysis on Multivariate, Time Series datasets using Shapelets
1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,
More informationData Mining and Machine Learning
Data Mining and Machine Learning Concept Learning and Version Spaces Introduction Concept Learning Generality Relations Refinement Operators Structured Hypothesis Spaces Simple algorithms Find-S Find-G
More informationSubcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network
Subcellular Localisation of Proteins in Living Cells Using a Genetic Algorithm and an Incremental Neural Network Marko Tscherepanow and Franz Kummert Applied Computer Science, Faculty of Technology, Bielefeld
More informationReview of Lecture 1. Across records. Within records. Classification, Clustering, Outlier detection. Associations
Review of Lecture 1 This course is about finding novel actionable patterns in data. We can divide data mining algorithms (and the patterns they find) into five groups Across records Classification, Clustering,
More informationStochastic Search: Part 2. Genetic Algorithms. Vincent A. Cicirello. Robotics Institute. Carnegie Mellon University
Stochastic Search: Part 2 Genetic Algorithms Vincent A. Cicirello Robotics Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 cicirello@ri.cmu.edu 1 The Genetic Algorithm (GA)
More informationNEUROEVOLUTION. Contents. Evolutionary Computation. Neuroevolution. Types of neuro-evolution algorithms
Contents Evolutionary Computation overview NEUROEVOLUTION Presenter: Vlad Chiriacescu Neuroevolution overview Issues in standard Evolutionary Computation NEAT method Complexification in competitive coevolution
More informationEvolutionary Computation: introduction
Evolutionary Computation: introduction Dirk Thierens Universiteit Utrecht The Netherlands Dirk Thierens (Universiteit Utrecht) EC Introduction 1 / 42 What? Evolutionary Computation Evolutionary Computation
More informationMachine Learning, Fall 2011: Homework 5
0-60 Machine Learning, Fall 0: Homework 5 Machine Learning Department Carnegie Mellon University Due:??? Instructions There are 3 questions on this assignment. Please submit your completed homework to
More informationCOMP9444: Neural Networks. Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization
: Neural Networks Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization 11s2 VC-dimension and PAC-learning 1 How good a classifier does a learner produce? Training error is the precentage
More informationEvolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia)
Evolutionary Computation DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia) andrea.roli@unibo.it Evolutionary Computation Inspiring principle: theory of natural selection Species face
More informationExamination Artificial Intelligence Module Intelligent Interaction Design December 2014
Examination Artificial Intelligence Module Intelligent Interaction Design December 2014 Introduction This exam is closed book, you may only use a simple calculator (addition, substraction, multiplication
More informationCS 380: ARTIFICIAL INTELLIGENCE MACHINE LEARNING. Santiago Ontañón
CS 380: ARTIFICIAL INTELLIGENCE MACHINE LEARNING Santiago Ontañón so367@drexel.edu Summary so far: Rational Agents Problem Solving Systematic Search: Uninformed Informed Local Search Adversarial Search
More informationCptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1
CptS 570 Machine Learning School of EECS Washington State University CptS 570 - Machine Learning 1 IEEE Expert, October 1996 CptS 570 - Machine Learning 2 Given sample S from all possible examples D Learner
More informationSearching the Subsumption Lattice by a Genetic Algorithm
Searching the Subsumption Lattice by a Genetic Algorithm Alireza Tamaddoni-Nezhad and Stephen H. Muggleton Department of Computer Science University of York, York, YO1 5DD, UK {alireza,stephen}@cs.york.ac.uk
More informationPAC Learning Introduction to Machine Learning. Matt Gormley Lecture 14 March 5, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University PAC Learning Matt Gormley Lecture 14 March 5, 2018 1 ML Big Picture Learning Paradigms:
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationA Note on Crossover with Interval Representations
A Note on Crossover with Interval Representations Christopher Stone and Larry Bull UWE Learning Classifier System Technical Report UWELCSG03-00 Faculty of Computing, Engineering and Mathematical Sciences
More informationA Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier
A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier Seiichi Ozawa, Shaoning Pang, and Nikola Kasabov Graduate School of Science and Technology, Kobe
More informationA new Approach to Drawing Conclusions from Data A Rough Set Perspective
Motto: Let the data speak for themselves R.A. Fisher A new Approach to Drawing Conclusions from Data A Rough et Perspective Zdzisław Pawlak Institute for Theoretical and Applied Informatics Polish Academy
More informationPattern Recognition Approaches to Solving Combinatorial Problems in Free Groups
Contemporary Mathematics Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups Robert M. Haralick, Alex D. Miasnikov, and Alexei G. Myasnikov Abstract. We review some basic methodologies
More informationData Mining und Maschinelles Lernen
Data Mining und Maschinelles Lernen Ensemble Methods Bias-Variance Trade-off Basic Idea of Ensembles Bagging Basic Algorithm Bagging with Costs Randomization Random Forests Boosting Stacking Error-Correcting
More informationCS 231A Section 1: Linear Algebra & Probability Review
CS 231A Section 1: Linear Algebra & Probability Review 1 Topics Support Vector Machines Boosting Viola-Jones face detector Linear Algebra Review Notation Operations & Properties Matrix Calculus Probability
More information