Use of Dempster-Shafer theory to combine classifiers which use different class boundaries

Size: px
Start display at page:

Download "Use of Dempster-Shafer theory to combine classifiers which use different class boundaries"

Transcription

1 Use of Dempster-Shafer theory to combine classifiers which use different class boundaries Mohammad Reza Ahmadzadeh and Maria Petrou Centre for Vision, Speech and Signal Processing, School of Electronics, Computing and Mathematics, University of Surrey, Guildford GU2 7XH, UK Tel: , Fax: Abstract In this paper we present the Dempster-Shafer theory as a framework within which the results of a Bayesian Network classifier and a fuzzy logic-based classifier are combined to produce a better final classification. We deal with the case when the two original classifiers use different classes for the outcome. The problem of different classes is solved by using a superset of finer classes which can be combined to produce classes according to either of the two classifiers. Within the Dempster-Shafer formalism not only the problem of different number of classes can be solved, but also the relative reliability of the classifiers can be taken into consideration. Keywords Classifier Combination, Dempster-Shafer Theory, Bayesian Networks, Fuzzy Logic, Expert Rules 1 Introduction It has been established recently that combining classifiers improves the classification accuracy for many problems. This has been established both theoretically, mainly within the framework of probability theory [7], and experimentally by many researchers. In addition, in the neural network field several approaches have been introduced to combine individual neural networks to improve the overall accuracy and performance. Currently with the School of Electronic Technology, Shiraz University, Shiraz, Iran, Tel: , Fax: , Ahmadzadeh@sanat.shirazu.ac.ir 1

2 Combining multiple neural network classifiers can be divided into two categories: Ensemble and modular [14], [15]. In the ensemble-based approaches, the number of output classes is the same for all classifiers. Each classifier has a complete solution for the problem, however combination of classifiers is used to improve the classification rate. On the other hand, in the modular approach, a problem is broken into several simple sub-problems. For example, a problem with 5 output classes can be changed into several subproblems with 2 output classes. Each sub-problem can be solved using a neural network. Combination of all classifiers provides a solution for the original problem. We can see that in this approach each classifier does not provide a solution for the problem, but all classifiers together are complementarily used to find the final classification. Combination of classifiers has also been investigated extensively when other types of classifier are used. Generally, classifiers can be combined at different levels: abstract level, ranking level and measurement level [19], [6]. In the abstract level only the top choice of each classifier is used for the purpose of combination. In the measurement level complete information for the outputs of the classifiers e.g. score for each possible output, is available and is used in the combination process. Although the combination at the abstract level uses the least information (only the top choice of the classifiers), it has been used frequently because all kinds of classifiers, such as statistical and syntactic, can be combined easily [19]. The ranking level approach may also be used to combine classifiers of any type at the expense of decision detail. It can be used not only to combine classifiers the outputs of which are either class rankings or confidence measures that can easily be converted into class rankings, but also classifiers that output single classes. The latter, however, is achieved by coarsening the the output of the other classifiers to comply with the classifier that outputs the least information [5]. Approaches that combine classifiers at the measurement level can combine any kind of classifiers that output measurements, but for the purpose of combination, these measurements should be translated into the same kind of measurement. For example, a classifier which supplies information at the output based on distances can not be directly combined with a classifier which outputs post-probability. Xu et al. in [19] used three methods of combining classifiers, all at the abstract level. They combined classifiers by using the Bayesian formalism, the voting principle and the Dempster-Shafer formalism. Their results on a case study showed that the Dempster-Shafer theory of evidence had the best results in comparison with the other methods. They used the recognition rate and substitution rate of classifiers to define 2

3 ) the mass functions. The mass function of the selected output (top choice) was defined from the recognition rate and the mass function of the other outputs (complement of the selected output) was defined from the substitution rate. If the summation of the recognition rate and the substitution rate was less than 100%, the remaining was called rejection rate and was assigned to the frame of discernment. This remaining indicated that the classifier was not able to decide, so it was interpreted as lack of information according to Dempster-Shafer theory. Rogova in [10] uses the Dempster-Shafer theory of evidence to combine neural network classifiers. All classifiers have the same number of outputs. So the frame of discernment,, is the same for all classifiers. represents the hypothesis that the output vector is of class. Let the th classifier be denoted by, the input vector by, and the output vector by! #", where the mean vector of the outputs of classifier ' be denoted by ( $$%. Further, let when the input is an element of the training set for class. A proximity measure can be defined using ( and. Rogova used a proximity measure, * % (, to define the mass functions. She defined various proximity measures in order to identify the best one. For any classifier and each class ) the proximity measure was defined to represent +, the support for hypothesis. Any evidence against or pro, -/. ),, was denoted by. Proximity measures which were considered as mass functions of the simple support functions were combined using a simplified version of Dempster s rule of combination. Having combined evidence from all classifiers, Dempster s rule of combination was used again to find the total confidence for each class. The class with the maximum confidence was singled out as the output of the classification. Rogova claimed that this method of combing classifiers could reduce misclassification by 15-30% compared with the best individual classifiers [10]. There are other classifier combination approaches in the literature, some of which were compared in [1]. The average [17], the weighted average, the Borda count [2], the fuzzy integral, the fuzzy connectives [9], the fuzzy templates and neural network approaches are among those which have been investigated in the literature. For an up-to-date overview of this research area, see for example the collection of papers in []. In all studies found in the literature so far, the classifiers combined are expected to use the same classes to classify the objects in question. In this paper we address the problem of different classes, which however span the same classification space. Some clarification is necessary here: When the classes of objects are 3

4 expressed by nouns, they are discrete and they are uniquely defined. Examples of such classes are chair, door, etc. However, there are problems where the classification of objects refers to some of their properties which may vary continuously. In such cases the defined classes are ad hoc quantisations of otherwise continuously varying properties. Such classes are for example very hot, hot, lukewarm, cold etc. In all such cases there is a hidden measurable quantity which takes continuous values and which characterises the state of the object. In the particular example, the measurable quantity is the temperature of the object, and the attribute according to which the object has to be classified is its thermal state. The division of this continuity of possible states into a small number of discrete classes can be done on the basis of the actual temperature value, using class boundaries that may be specified arbitrarily. Two different experts (classifiers) may specify different class boundaries. It is this type of problem we address here, with the help of the Dempster-Shafer theory of evidence, where the results of the two classifiers are considered as items of evidence in support of a certain proposition. The problem of different classes is solved by using a superset of finer classes which can be combined to produce classes according to either of the two classifiers. The nearest research to our work is that of using error correcting codes (ECC) [3] for classification. An ECC uses more characteristics than there are necessary to classify something and then it maps the superset of characteristics to the minimum meaningful number that are needed for the classification. One may see each characteristic as defining a different class and consider that ECC maps a larger number of classes to a smaller one (the correct ones). Our method differs in several respects: 1) The classes used by either of the classifiers are both legitimate sets of classes. 2) Our method can be used to refine/improve the performance of either classifier. 3) ECC uses classifiers that have as output yes/no answers, while the approach we use here comes up with probabilities assigned to each possible class. Our method is demonstrated in conjunction with the problem of predicting the risk of soil erosion of burned forests in the Mediterranean region using data concerning relevant factors like soil depth, ground slope and rock permeability. This problem has been solved in the past using Pearl-Bayes networks [16] and Fuzzy Logic [12], [11]. The results of these classifiers were available to us and they are combined to produce a more reliable classification. 4

5 2 Data Description Soil erosion depends on three variables: slope, soil depth and rock permeability. Other factors that may influence soil erosion are not taken into account as they were uniform in the area of study which our data refer to. Geophysical data are available from 39 sites of four areas in Greece. For each of these sites the classification by a human expert to one of five possible classes of soil erosion is also available. Each of the problem variables takes values from a small set of possible classes. Stassopoulou et al. [16] implemented a Pearl-Bayes network with which they solved the problem of combining the values of the attributes, alongside the uncertainties associated with them, in order to infer the probability with which a site belonged to one of the possible classes of risk of soil erosion. The use of a Pearl-Bayes network involved the use of conditional probability functions. For the case when the combined attributes and inferred conclusions are discrete valued quantities, these conditional probabilities are matrices. In the particular case, as three attributes were combined to assess the risk of soil erosion, if each variable involved could take 0 possible values, the matrix ought to have been So, for 0 76, there should be 625 elements of the matrix, each expressing the probability of the site to belong to a certain class of soil erosion, given that the site attributes have certain combination of classes. The calculation of such a large number of probabilities, however, required the availability of a large number of data. In research problems one very seldomly has at one s disposal enough relevant data for such estimation. To reduce the severity of the problem, Stassopoulou et al. quantised all variables of the problem into three classes each, thus having to compute only 1 conditional probability values. Their results were quite satisfactory: They obtained consistent results on the training set for 2 out of the 30 training sites, and hardening their conclusions produced agreement with the expert in 7 out of the 9 test sites. However, in spite of their accuracy, these results used gross classes, as each variable was quantised only into one of 3 possible classes. Sasikala et al. [11], solved the same problem, using the same data, but as no numerical restriction existed, their results classified the risk of soil erosion into one of five possible classes, the same ones used by the expert who had made the assessment in the first place. Sasikala et al., in order to solve this problem developed a new fuzzy methodology, which involved a training stage: weights were used for the membership functions to reflect the relative importance of the combined attributes, and many different combination rules were tried. The system was trained for the selection of the best set of weights and the best combination rule. Upon hardening the final classification, they could have consistency in the training 5

6 D _??? D Z, data in 1 out of the 30 sites and they could predict correctly the class of the test sites in 5 out of the 9 cases. However, the use of weights and a variety of combination rules, produced a blunt decision system: in some cases more than one possible classes had equally high membership functions. The idea we propose here is to combine the results of the accurate probabilistic classifier, which uses gross classes, with the results of the blunt fuzzy classifier, which uses finer classes, in order to obtain a final classification which will be more accurate and less blunt. 3 Dempster-Shafer Theory The theory of evidence was introduced by Glean Shafer in 1976 as a mathematical framework for the representation of uncertainty. Let. The mass function follows [13], [1]. for all non empty sets?ut :9!%@?A where and be two mass functions on the same frame of discernment, which is called orthogonal summation of <; and >= is defined as ; ;CBED F GVHIKJ'W F G:HIKJML %RN!P %ONEP %@QX %RQS (1) is considered as a normalisation factor and is needed to make sure that no mass is assigned to the empty set. In addition it is a measure of conflict between the two masses. If DY. conflict. If D ; it means that there is no ; the combination of!; and 5= does not exist and they are totally or flatly contradictory. For illustrating purposes it is convenient to represent this rule of combination graphically, as shown in figure 1. Let?, and Z Z\[ be focal elements of mass functions respectively. Along the horizontal side of a unit square we show the mass functions of all elements of The width of each strip is proportional to the value of its corresponding mass function. The mass functions of are shown along the vertical side of the same square. The area of intersection of strips %@Z [ (dashed area) represents the amount of mass that is assigned to? C9 combination rule, E%@_`,] Z [. It is possible that for some - and a we have?,^] %@? and. and Z [. According to Dempster s %@_` is proportional to the sum of the areas of all rectangles where,] Z [ cb. The mass functions have to be scaled in order to make sure that the sum of a mass function over all subsets of is 1. This is done with the help of D in equation (1). If? set and we say that and, ] Z\[ db for all - and a, all mass of the combination goes to the empty are not combinable or are totally contradictory. 6

7 _, eë e %R? eë e //////// %@ZC[ //////// ////////. Figure 1: Combination of mass functions 4 The Proposed System of Classifier Combination In this paper we use the Dempster-Shafer theory to combine the items of evidence that come from the Bayesian network of Stassopoulou et al. and the fuzzy logic classifier of Sasikala et al. One of the conditions to be able to use the Dempster-Shafer theory is that all sources should have the same frame of discernment [13], [4]. In our case this is not true, as for example, risk of soil erosion is classified into 3 classes, which we denote by? denote by Z,, ZAh, in the fuzzy logic method.,?,?gf, in the Bayesian network method, and into 5 classes, which we To be able to use the Dempster-Shafer theory in this application, we look for a definition of a frame of discernment in which both methods can be defined. Since both methods span the same classification space, we quantise the classification space into 15 classes, _ to _ h. These classes can be expressed in both methods because 15 can be divided by 3 and 5. In other words, the union, for example, of the first 5 new classes i.e. _ : V _:h union of the first 3 new classes i.e. _ is the same as the first class of the Bayesian network method, i.e.? _if is the same as the first class, i.e. Z. Also the, of the fuzzy logic method. Figure 2 shows schematically the idea of defining this superset of classes. The next step is defining the mass functions from two classifiers. We interpret the beliefs of the Bayesian network system as mass functions in the Dempster-Shafer theory. Since the output measurements of a Bayesian network are in the form of probabilities, no further conversion is needed to use them as mass functions. However, the membership grades of classes in the fuzzy system, although in the range j k they do not sum up to 1. Therefore we cannot interpret them as mass functions directly. Instead, we use them in order to distribute the unit mass, proportionally, to the corresponding classes. 7 ;l,

8 A1 A 2 A 3 B 1 B 2 B 3 B 4 B 5 C1... C 15 Figure 2: Definition of the superset of classes. Before using the defined mass functions in Dempster s combination rule, another factor that should be taken into consideration is the relative reliability of the two classifiers. If we have the recognition rate, substitution rate (error rate) and rejection rate of a classifier, its reliability can be defined as [19]: nmo -qprs- o -qtp nmuvxw -PtP- v ; k+k^y Bnm a mxu tp- v (2) If a classifier does not include a rejection option, like the Bayesian classifier of Stassopoulou et al. [16], its reliability is the same as its recognition rate. So, we are going to use as reliability of the Bayesian classifier its recognition rate. The fuzzy logic classifier, however, was based heavily on the use of individual production rules, which themselves might be treated as a collection of individual classifiers. One does not necessarily expect that all rules used are equally reliable; indeed, some of them may even be wrong. To assign, therefore, an overall reliability factor to the fuzzy classifier would be equivalent to ignoring the peculiarities of the individual classifiers this is a collection of. We decided instead, to examine, with the help of a training phase, the reliability of the individual firing rules. They are these individual reliability factors that are used to moderate the mass functions of the fuzzy classifier. In the Dempster-Shafer theory we can interpret unreliability of a source as lack of information. So, after we scale down the mass functions which we have already defined for each classifier, by taking into consideration the reliability of the classifier, we assign the remaining mass to the frame of discernment as lack of information. In figure 3 the mass functions derived from the Bayesian network and the fuzzy logic system after considering the reliability of the classifiers, are denoted by and respectively. The combination of

9 D z`{} ~ { z`{} ~ ƒ z`{} ~K z {} _ z ƒ { Š Š Œ Œ Š Š z ƒ ƒ Œ z ƒ Œ Œ z ƒ Œ Š Š Š z ƒ Œ Œ Š { Š { Š { Š { Š Ž Ž ˆ z` Š {} ƒ z Š { ƒ ˆ z Š$ ŠŽ z` Š$ z Š$ ŠŽ Š$ ˆ z Š Š$ Š$ z Š Š$ Š$ ˆM z Š {R z` Š {q{ {@ƒ z Š {R {q{} {@ƒ ˆKŽ z` Š {R Ž z Š {R Ž znƒ z` ŠK{ z Š ŠK{R z ŠK{q{ ŠK{ z` Figure 3: Combination of mass functions mass functions is denoted by. Note that the square area denoted by, for example,!% _ is equal to %@? A1E %RZ. This value is used in Dempster s rule of combination given by equation 1 in order to assign mass to _. As it can be seen, in fifteen cases, the mass functions which resulted from the combination of the two sources can be assigned to non empty sets. Here the normalisation factor, D, is: %@? P %R?gfq %RZ Mš %RZAfš %@? f P %R? q %@Z šœ %RZgžšŸ %R? f q %R? q %RZ š %RZAhš %R? %R? q q %RZ f š (3) %RZAhx For example, we have: E%@_ + _:fx %R? q ;CB!D %RZ (4) Although we have classified the risk of soil erosion into 15 classes, we would like to have the result in 5 classes as used by the expert and by the fuzzy logic system. Thus, we calculate the belief function of the classes of interest, by using the mass functions of the focal elements. So, after scaling all mass functions which are assigned to non empty subsets, the summation of masses of the classes in each row will be the belief function of the corresponding class. For example, summation of E%@_\ž E%@_iž _:h _i in the second row will be assigned as r mxo %@_Vž out of the 5 possible classes, i.e. _:h _:hx,!% _i and _: x which is the belief of the second class r mxo %RZ r mo %@_ ž _ h _!% _ ž _ h 'š E%@_ š E%@_ ž _ h _ Figure 4 shows schematically the proposed combination system. 9

10 » Ä Ä A Slope Bayesian Network Mass Function Gen. Soil depth Dempster s Rock per. Fuzzy Logic System Mass Function Gen. B Combination Rule Reliability of Classifiers Figure 4: Proposed combination of classifiers system to assess the risk of soil erosion. 5 Experimental Results If we denote the output beliefs of the Bayesian network by `, ª and its recognition rate by «, we used: O±q²³s±q O±qṔµ «º¹» ¼¼¾½ R gàp Á R gàp VÂà ¼+¼ ±» Ä ÅÆÄÇ with «ÉÈÈ ½ ÈÊ. To deal with the reliability of the fuzzy classifier, we multiplied with weights, ¼SËÍÌ ÀºÎ» Ä ÅÄ ½½ ½ Ä Ï, the different mass functions which resulted from different expert rules used by it. We used 30 training sites to identify the best weights which would produce the best results. It is worth mentioning that we used exhaustive space search to find the best weights. However, in every set of weights we fixed one of the weights to be 1 in order to make the space search smaller, and because this way the weights measured the relative importance of the various rules. We found that the best results could be obtained when the mass functions of classes and Ð were scaled down by Ñ This is a very interesting outcome as it indicates that perhaps less emphasis should be placed on rules that lead to classes other than the two extremes and the middle. We speculate that perhaps people find easy to classify things in classes like low, medium and high, or good, average and bad etc. It is more difficult to ask people to classify things in classes like very low, low, medium, high and very high, or very good, good, average, bad and very bad. It seems that heuristics devised by people to yield classification in classes inserted between the 3 gross ones, may not be as reliable as rules that classify into the clear cut and well defined classes of two extremes and a medium. ± 10

11 , % % Z? F f F h,,, After the reliability of each classifier was taken into consideration, the sum of its mass functions was not 1. The difference of the sum from 1 was assigned to the frame of discernment which is interpreted as the lack of information. For example, for the fuzzy classifier where Ò %RZ ZAf are the weights identified during the training phase. For the Bayesian classifier %@? Zgž? f ZAhx ;CB ;CB, J, J KÒ Z(`Ó %@? ; k+k By using the maximum belief function criterion in the decision making step, 6 out of the 9 testing sites were correctly classified and the other 3 sites were classified in the next class from that assigned by the expert. This should be compared with the 5 sites which were correctly classified by the fuzzy classifier alone. 6 Discussion and Conclusions In classification problems where the classes used represent different grades of the same attribute, it is possible to have different divisions into classes used by different classifiers. A similar situation may arise when the individual classifiers are unsupervised and determine the data classes automatically. In such cases it is not possible to combine the results of the different classifiers in a straightforward way. We proposed here that one may handle such problems within the framework of Dempster-Shafer theory. The Dempster- Shafer theory as a classifier combination method allows one to deal with the different number of classes used by the different classifiers, because, unlike Bayesian theory, it assigns probabilities to sets of possible hypotheses, not just to individual hypotheses. In addition, it allows one to take into consideration the reliability of the classifiers in the process of mass definition. We demonstrated our ideas using an example problem of prediction of soil erosion. Within the limitations of this problem, we showed that not only the accuracy of the individual classifiers was improved but also that a finer set of output classes could be obtained. Although our results are too limited and their statistical significance can not be estimated due to the limitations of our dataset, this should not degrade the proposed methodology which is generic and applicable to any situation where the classes defined by the different classifiers are different. This stems from the ability of the Dempster-Shafer theory to assign probabilities to sets of possible classes and not just to individual classes. 11 qô L %@Z

12 Acknowledgements Mohammad Reza Ahmadzadeh was on leave from the University of Shiraz, Iran, when this work was carried out as part of his PhD thesis. He is grateful to the Ministry of Science, Research and Technology of Iran for its financial support during his research studies. References [1] J. A. Barnett. Computational methods for a mathematical theory of evidence. In Proceedings of the 7th International Joint Conference on Artificial Intelligence., Vancouver, BC, Ca, volume 2, pages 6 75, Aug [2] C. Black. The Theory of Committees and Elections. Cambridge University Press, [3] T. G. Dietterich and G. Bakiri. Solving multiclass learning problems via error-correcting output codes. Journal of Artificial Intelligence Research, 2:263 26, [4] J. W. Guan and D. A. Bell. Evidence Theory and its Applications. ELSEVIER SCIENCE PUBLISHER B.V., [5] T. K. Ho, J. J. Hull and S. N. Srihari. Decision Combination in Multiple Classifier Systems. IEEE Transactions on Pattern Analysis and Machine Intelligence, 16(1):66 75, [6] H. Kang, K. Kim, and J. H. Kim. A framework for probabilistic combination of multiple classifiers at an abstract level. Engineering Applications of Artificial Intelligence, 10(4):379 35, [7] J. Kittler. Combining classifiers: A theoretical framework. Pattern Analysis and Applications, 1:1 27, Jan 199. [] J. Kittler and F. Roli (Eds). Multiple Classifier Systems. Springer LNCS 2096, ISBN , [9] L. Kuncheva. An application of owa operators to the aggregation of multiple classification decisions. In R. Yager and J. Kacprzyk, editors, The Ordered Weighted Averaging Operators. Theory and Applications. Kluwer Academic Publishewrs, [10] G. Rogova. Combining the result of several neural network classifiers. Neural Networks, 7(5):777 71, [11] K. R. Sasikala and M. Petrou. Generalised fuzzy aggregation in estimating the risk of desertification of a burned forest. Fuzzy Sets and Systems, 11(1): , February [12] K. R. Sasikala, M. Petrou, and J. Kittler. Fuzzy classification with a GIS as an aid to decision making. EARSeL Advances in remote sensing, 4(4):97 105, November [13] G. Shafer. A Mathematical Theory of Evidence. Princeton University Press, [14] A. J. C. Sharkey. On combining artificial neural nets. Connection Science, (3/4): ,

13 [15] A. J. C. Sharkey. Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems. Springer-Verlag, 199. [16] A. Stassopoulou, M. Petrou, and J. Kittler. Application of a Bayesian network in a GIS based decision making system. Int. J. Geographical Information Science, 12(1):23 45, 199. [17] M. Taniguchi and V. Tresp. Averaging regularized estimation. Neural Computation, 9: , [1] A. Verikas, A. Lipnickas, K. Malmqvist, M.Bacauskiene, and A. Gelzinis. Soft combination of neural classifiers: A comparative study. Pattern Recognition Letters, 20(4): , [19] L. Xu, A. Krzyzak, and C. Y. Suen. Methods of combining multiple classifiers and their applications to handwriting recognition. IEEE Transactions on Systems, Man and Cybernetics, 22(3):41 435,

Selection of Classifiers based on Multiple Classifier Behaviour

Selection of Classifiers based on Multiple Classifier Behaviour Selection of Classifiers based on Multiple Classifier Behaviour Giorgio Giacinto, Fabio Roli, and Giorgio Fumera Dept. of Electrical and Electronic Eng. - University of Cagliari Piazza d Armi, 09123 Cagliari,

More information

Uncertainty and Rules

Uncertainty and Rules Uncertainty and Rules We have already seen that expert systems can operate within the realm of uncertainty. There are several sources of uncertainty in rules: Uncertainty related to individual rules Uncertainty

More information

Data Dependence in Combining Classifiers

Data Dependence in Combining Classifiers in Combining Classifiers Mohamed Kamel, Nayer Wanas Pattern Analysis and Machine Intelligence Lab University of Waterloo CANADA ! Dependence! Dependence Architecture! Algorithm Outline Pattern Recognition

More information

Handling imprecise and uncertain class labels in classification and clustering

Handling imprecise and uncertain class labels in classification and clustering Handling imprecise and uncertain class labels in classification and clustering Thierry Denœux 1 1 Université de Technologie de Compiègne HEUDIASYC (UMR CNRS 6599) COST Action IC 0702 Working group C, Mallorca,

More information

Data Mining und Maschinelles Lernen

Data Mining und Maschinelles Lernen Data Mining und Maschinelles Lernen Ensemble Methods Bias-Variance Trade-off Basic Idea of Ensembles Bagging Basic Algorithm Bagging with Costs Randomization Random Forests Boosting Stacking Error-Correcting

More information

Combination Methods for Ensembles of Multilayer Feedforward 1

Combination Methods for Ensembles of Multilayer Feedforward 1 Combination Methods for Ensembles of Multilayer Feedforward 1 JOAQUÍN TORRES-SOSPEDRA MERCEDES FERNÁNDEZ-REDONDO CARLOS HERNÁNDEZ-ESPINOSA Dept. de Ingeniería y Ciencia de los Computadores Universidad

More information

The internal conflict of a belief function

The internal conflict of a belief function The internal conflict of a belief function Johan Schubert 1 Abstract In this paper we define and derive an internal conflict of a belief function We decompose the belief function in question into a set

More information

Managing Decomposed Belief Functions

Managing Decomposed Belief Functions Managing Decomposed Belief Functions Johan Schubert Department of Decision Support Systems, Division of Command and Control Systems, Swedish Defence Research Agency, SE-164 90 Stockholm, Sweden schubert@foi.se

More information

A NEW CLASS OF FUSION RULES BASED ON T-CONORM AND T-NORM FUZZY OPERATORS

A NEW CLASS OF FUSION RULES BASED ON T-CONORM AND T-NORM FUZZY OPERATORS A NEW CLASS OF FUSION RULES BASED ON T-CONORM AND T-NORM FUZZY OPERATORS Albena TCHAMOVA, Jean DEZERT and Florentin SMARANDACHE Abstract: In this paper a particular combination rule based on specified

More information

Naive Bayesian Rough Sets

Naive Bayesian Rough Sets Naive Bayesian Rough Sets Yiyu Yao and Bing Zhou Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 {yyao,zhou200b}@cs.uregina.ca Abstract. A naive Bayesian classifier

More information

A Simple Implementation of the Stochastic Discrimination for Pattern Recognition

A Simple Implementation of the Stochastic Discrimination for Pattern Recognition A Simple Implementation of the Stochastic Discrimination for Pattern Recognition Dechang Chen 1 and Xiuzhen Cheng 2 1 University of Wisconsin Green Bay, Green Bay, WI 54311, USA chend@uwgb.edu 2 University

More information

Reasoning with Uncertainty

Reasoning with Uncertainty Reasoning with Uncertainty Representing Uncertainty Manfred Huber 2005 1 Reasoning with Uncertainty The goal of reasoning is usually to: Determine the state of the world Determine what actions to take

More information

A novel k-nn approach for data with uncertain attribute values

A novel k-nn approach for data with uncertain attribute values A novel -NN approach for data with uncertain attribute values Asma Trabelsi 1,2, Zied Elouedi 1, and Eric Lefevre 2 1 Université de Tunis, Institut Supérieur de Gestion de Tunis, LARODEC, Tunisia trabelsyasma@gmail.com,zied.elouedi@gmx.fr

More information

arxiv: v1 [cs.ai] 28 Oct 2013

arxiv: v1 [cs.ai] 28 Oct 2013 Ranking basic belief assignments in decision making under uncertain environment arxiv:30.7442v [cs.ai] 28 Oct 203 Yuxian Du a, Shiyu Chen a, Yong Hu b, Felix T.S. Chan c, Sankaran Mahadevan d, Yong Deng

More information

Where are we? Knowledge Engineering Semester 2, Reasoning under Uncertainty. Probabilistic Reasoning

Where are we? Knowledge Engineering Semester 2, Reasoning under Uncertainty. Probabilistic Reasoning Knowledge Engineering Semester 2, 2004-05 Michael Rovatsos mrovatso@inf.ed.ac.uk Lecture 8 Dealing with Uncertainty 8th ebruary 2005 Where are we? Last time... Model-based reasoning oday... pproaches to

More information

Today s s lecture. Lecture 16: Uncertainty - 6. Dempster-Shafer Theory. Alternative Models of Dealing with Uncertainty Information/Evidence

Today s s lecture. Lecture 16: Uncertainty - 6. Dempster-Shafer Theory. Alternative Models of Dealing with Uncertainty Information/Evidence Today s s lecture Lecture 6: Uncertainty - 6 Alternative Models of Dealing with Uncertainty Information/Evidence Dempster-Shaffer Theory of Evidence Victor Lesser CMPSCI 683 Fall 24 Fuzzy logic Logical

More information

Previous Accomplishments. Focus of Research Iona College. Focus of Research Iona College. Publication List Iona College. Journals

Previous Accomplishments. Focus of Research Iona College. Focus of Research Iona College. Publication List Iona College. Journals Network-based Hard/Soft Information Fusion: Soft Information and its Fusion Ronald R. Yager, Tel. 212 249 2047, E-Mail: yager@panix.com Objectives: Support development of hard/soft information fusion Develop

More information

Scientific/Technical Approach

Scientific/Technical Approach Network based Hard/Soft Information Fusion: Soft Information and its Fusion Ronald R. Yager, Tel. 212 249 2047, E Mail: yager@panix.com Objectives: Support development of hard/soft information fusion Develop

More information

The Semi-Pascal Triangle of Maximum Deng Entropy

The Semi-Pascal Triangle of Maximum Deng Entropy The Semi-Pascal Triangle of Maximum Deng Entropy Xiaozhuan Gao a, Yong Deng a, a Institute of Fundamental and Frontier Science, University of Electronic Science and Technology of China, Chengdu, 610054,

More information

A generic framework for resolving the conict in the combination of belief structures E. Lefevre PSI, Universite/INSA de Rouen Place Emile Blondel, BP

A generic framework for resolving the conict in the combination of belief structures E. Lefevre PSI, Universite/INSA de Rouen Place Emile Blondel, BP A generic framework for resolving the conict in the combination of belief structures E. Lefevre PSI, Universite/INSA de Rouen Place Emile Blondel, BP 08 76131 Mont-Saint-Aignan Cedex, France Eric.Lefevre@insa-rouen.fr

More information

Risk Assessment of E-Commerce Projects Using Evidential Reasoning

Risk Assessment of E-Commerce Projects Using Evidential Reasoning Risk Assessment of E-Commerce Projects Using Evidential Reasoning Rashid Hafeez Khokhar, David A. Bell, Jiwen Guan, and QingXiang Wu The School of Electronics, Electrical Engineering and Computer Science

More information

Solving Classification Problems By Knowledge Sets

Solving Classification Problems By Knowledge Sets Solving Classification Problems By Knowledge Sets Marcin Orchel a, a Department of Computer Science, AGH University of Science and Technology, Al. A. Mickiewicza 30, 30-059 Kraków, Poland Abstract We propose

More information

Multicomponent DS Fusion Approach for Waveform EKG Detection

Multicomponent DS Fusion Approach for Waveform EKG Detection Multicomponent DS Fusion Approach for Waveform EKG Detection Nicholas Napoli University of Virginia njn5fg@virginia.edu August 10, 2013 Nicholas Napoli (UVa) Multicomponent EKG Fusion August 10, 2013 1

More information

Measure divergence degree of basic probability assignment based on Deng relative entropy

Measure divergence degree of basic probability assignment based on Deng relative entropy Measure divergence degree of basic probability assignment based on Deng relative entropy Liguo Fei a, Yong Deng a,b,c, a School of Computer and Information Science, Southwest University, Chongqing, 400715,

More information

A Novel Rejection Measurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis

A Novel Rejection Measurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis 009 0th International Conference on Document Analysis and Recognition A Novel Rejection easurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis Chun Lei He Louisa Lam Ching

More information

Imprecise Probability

Imprecise Probability Imprecise Probability Alexander Karlsson University of Skövde School of Humanities and Informatics alexander.karlsson@his.se 6th October 2006 0 D W 0 L 0 Introduction The term imprecise probability refers

More information

COMP9444: Neural Networks. Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization

COMP9444: Neural Networks. Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization : Neural Networks Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization 11s2 VC-dimension and PAC-learning 1 How good a classifier does a learner produce? Training error is the precentage

More information

Machine Learning and Deep Learning! Vincent Lepetit!

Machine Learning and Deep Learning! Vincent Lepetit! Machine Learning and Deep Learning!! Vincent Lepetit! 1! What is Machine Learning?! 2! Hand-Written Digit Recognition! 2 9 3! Hand-Written Digit Recognition! Formalization! 0 1 x = @ A Images are 28x28

More information

Conditional Deng Entropy, Joint Deng Entropy and Generalized Mutual Information

Conditional Deng Entropy, Joint Deng Entropy and Generalized Mutual Information Conditional Deng Entropy, Joint Deng Entropy and Generalized Mutual Information Haoyang Zheng a, Yong Deng a,b, a School of Computer and Information Science, Southwest University, Chongqing 400715, China

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

CMPT Machine Learning. Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th

CMPT Machine Learning. Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th CMPT 882 - Machine Learning Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th Stephen Fagan sfagan@sfu.ca Overview: Introduction - Who was Bayes? - Bayesian Statistics Versus Classical Statistics

More information

Explaining Results of Neural Networks by Contextual Importance and Utility

Explaining Results of Neural Networks by Contextual Importance and Utility Explaining Results of Neural Networks by Contextual Importance and Utility Kary FRÄMLING Dep. SIMADE, Ecole des Mines, 158 cours Fauriel, 42023 Saint-Etienne Cedex 2, FRANCE framling@emse.fr, tel.: +33-77.42.66.09

More information

SYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS

SYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS SYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS Hans-Jürgen Winkler ABSTRACT In this paper an efficient on-line recognition system for handwritten mathematical formulas is proposed. After formula

More information

Sequential adaptive combination of unreliable sources of evidence

Sequential adaptive combination of unreliable sources of evidence Sequential adaptive combination of unreliable sources of evidence Zhun-ga Liu, Quan Pan, Yong-mei Cheng School of Automation Northwestern Polytechnical University Xi an, China Email: liuzhunga@gmail.com

More information

Lecture 2. Judging the Performance of Classifiers. Nitin R. Patel

Lecture 2. Judging the Performance of Classifiers. Nitin R. Patel Lecture 2 Judging the Performance of Classifiers Nitin R. Patel 1 In this note we will examine the question of how to udge the usefulness of a classifier and how to compare different classifiers. Not only

More information

Rough Set Model Selection for Practical Decision Making

Rough Set Model Selection for Practical Decision Making Rough Set Model Selection for Practical Decision Making Joseph P. Herbert JingTao Yao Department of Computer Science University of Regina Regina, Saskatchewan, Canada, S4S 0A2 {herbertj, jtyao}@cs.uregina.ca

More information

Fuzzy Systems. Possibility Theory.

Fuzzy Systems. Possibility Theory. Fuzzy Systems Possibility Theory Rudolf Kruse Christian Moewes {kruse,cmoewes}@iws.cs.uni-magdeburg.de Otto-von-Guericke University of Magdeburg Faculty of Computer Science Department of Knowledge Processing

More information

arxiv: v1 [cs.cv] 11 Jun 2008

arxiv: v1 [cs.cv] 11 Jun 2008 HUMAN EXPERTS FUSION FOR IMAGE CLASSIFICATION Arnaud MARTIN and Christophe OSSWALD arxiv:0806.1798v1 [cs.cv] 11 Jun 2008 Abstract In image classification, merging the opinion of several human experts is

More information

CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE

CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE CHAPTER 3. THE IMPERFECT CUMULATIVE SCALE 3.1 Model Violations If a set of items does not form a perfect Guttman scale but contains a few wrong responses, we do not necessarily need to discard it. A wrong

More information

Belief Update in CLG Bayesian Networks With Lazy Propagation

Belief Update in CLG Bayesian Networks With Lazy Propagation Belief Update in CLG Bayesian Networks With Lazy Propagation Anders L Madsen HUGIN Expert A/S Gasværksvej 5 9000 Aalborg, Denmark Anders.L.Madsen@hugin.com Abstract In recent years Bayesian networks (BNs)

More information

An Analytical Comparison between Bayes Point Machines and Support Vector Machines

An Analytical Comparison between Bayes Point Machines and Support Vector Machines An Analytical Comparison between Bayes Point Machines and Support Vector Machines Ashish Kapoor Massachusetts Institute of Technology Cambridge, MA 02139 kapoor@mit.edu Abstract This paper analyzes the

More information

COMPARING PERFORMANCE OF NEURAL NETWORKS RECOGNIZING MACHINE GENERATED CHARACTERS

COMPARING PERFORMANCE OF NEURAL NETWORKS RECOGNIZING MACHINE GENERATED CHARACTERS Proceedings of the First Southern Symposium on Computing The University of Southern Mississippi, December 4-5, 1998 COMPARING PERFORMANCE OF NEURAL NETWORKS RECOGNIZING MACHINE GENERATED CHARACTERS SEAN

More information

Ensemble Methods and Random Forests

Ensemble Methods and Random Forests Ensemble Methods and Random Forests Vaishnavi S May 2017 1 Introduction We have seen various analysis for classification and regression in the course. One of the common methods to reduce the generalization

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

Divergence measure of intuitionistic fuzzy sets

Divergence measure of intuitionistic fuzzy sets Divergence measure of intuitionistic fuzzy sets Fuyuan Xiao a, a School of Computer and Information Science, Southwest University, Chongqing, 400715, China Abstract As a generation of fuzzy sets, the intuitionistic

More information

Bayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington

Bayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington Bayesian Classifiers and Probability Estimation Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington 1 Data Space Suppose that we have a classification problem The

More information

Basic Probabilistic Reasoning SEG

Basic Probabilistic Reasoning SEG Basic Probabilistic Reasoning SEG 7450 1 Introduction Reasoning under uncertainty using probability theory Dealing with uncertainty is one of the main advantages of an expert system over a simple decision

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

Bagging and Boosting for the Nearest Mean Classifier: Effects of Sample Size on Diversity and Accuracy

Bagging and Boosting for the Nearest Mean Classifier: Effects of Sample Size on Diversity and Accuracy and for the Nearest Mean Classifier: Effects of Sample Size on Diversity and Accuracy Marina Skurichina, Liudmila I. Kuncheva 2 and Robert P.W. Duin Pattern Recognition Group, Department of Applied Physics,

More information

A Static Evidential Network for Context Reasoning in Home-Based Care Hyun Lee, Member, IEEE, Jae Sung Choi, and Ramez Elmasri, Member, IEEE

A Static Evidential Network for Context Reasoning in Home-Based Care Hyun Lee, Member, IEEE, Jae Sung Choi, and Ramez Elmasri, Member, IEEE 1232 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART A: SYSTEMS AND HUMANS, VOL. 40, NO. 6, NOVEMBER 2010 A Static Evidential Network for Context Reasoning in Home-Based Care Hyun Lee, Member,

More information

Reasoning Under Uncertainty

Reasoning Under Uncertainty Reasoning Under Uncertainty Chapter 14&15 Part Kostas (1) Certainty Kontogiannis Factors E&CE 457 Objectives This unit aims to investigate techniques that allow for an algorithmic process to deduce new

More information

Multisource Image Fusion Algorithm Based On A New Evidential Reasoning Approach

Multisource Image Fusion Algorithm Based On A New Evidential Reasoning Approach 1 Multisource Image Fusion Algorithm Based On A New Evidential Reasoning Approach Mickaël Germain 1, Matthieu Voorons 1, Jean-Marc Boucher 2, Goze B. Bénié 1 et Éric Beaudry 1 (1) Centre d applications

More information

Pattern recognition. "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher

Pattern recognition. To understand is to perceive patterns Sir Isaiah Berlin, Russian philosopher Pattern recognition "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher The more relevant patterns at your disposal, the better your decisions will be. This is hopeful news to

More information

Linear Combiners for Fusion of Pattern Classifiers

Linear Combiners for Fusion of Pattern Classifiers International Summer School on eural ets duardo R. Caianiello" SMBL MTHODS FOR LARIG MACHIS Vietri sul Mare(Salerno) ITALY 22-28 September 2002 Linear Combiners for Fusion of Pattern Classifiers Lecturer

More information

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Article from. Predictive Analytics and Futurism. July 2016 Issue 13 Article from Predictive Analytics and Futurism July 2016 Issue 13 Regression and Classification: A Deeper Look By Jeff Heaton Classification and regression are the two most common forms of models fitted

More information

Analyzing the degree of conflict among belief functions.

Analyzing the degree of conflict among belief functions. Analyzing the degree of conflict among belief functions. Liu, W. 2006). Analyzing the degree of conflict among belief functions. Artificial Intelligence, 17011)11), 909-924. DOI: 10.1016/j.artint.2006.05.002

More information

Generalised Fuzzy Aggregation in Estimating the Risk of Desertification of a burned forest

Generalised Fuzzy Aggregation in Estimating the Risk of Desertification of a burned forest Generalised Fuzzy Aggregation in Estimating the Risk of Desertification of a burned forest K. R. Sasikala and M. Petrou School of Electronic Engineering, Information Technology and Mathematics University

More information

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH Hoang Trang 1, Tran Hoang Loc 1 1 Ho Chi Minh City University of Technology-VNU HCM, Ho Chi

More information

Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric?

Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric? Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric? Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/

More information

Rough operations on Boolean algebras

Rough operations on Boolean algebras Rough operations on Boolean algebras Guilin Qi and Weiru Liu School of Computer Science, Queen s University Belfast Belfast, BT7 1NN, UK Abstract In this paper, we introduce two pairs of rough operations

More information

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1 EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle

More information

A Large Deviation Bound for the Area Under the ROC Curve

A Large Deviation Bound for the Area Under the ROC Curve A Large Deviation Bound for the Area Under the ROC Curve Shivani Agarwal, Thore Graepel, Ralf Herbrich and Dan Roth Dept. of Computer Science University of Illinois Urbana, IL 680, USA {sagarwal,danr}@cs.uiuc.edu

More information

The Lady Tasting Tea. How to deal with multiple testing. Need to explore many models. More Predictive Modeling

The Lady Tasting Tea. How to deal with multiple testing. Need to explore many models. More Predictive Modeling The Lady Tasting Tea More Predictive Modeling R. A. Fisher & the Lady B. Muriel Bristol claimed she prefers tea added to milk rather than milk added to tea Fisher was skeptical that she could distinguish

More information

Evaluation requires to define performance measures to be optimized

Evaluation requires to define performance measures to be optimized Evaluation Basic concepts Evaluation requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain (generalization error) approximation

More information

Introduction to Machine Learning. Introduction to ML - TAU 2016/7 1

Introduction to Machine Learning. Introduction to ML - TAU 2016/7 1 Introduction to Machine Learning Introduction to ML - TAU 2016/7 1 Course Administration Lecturers: Amir Globerson (gamir@post.tau.ac.il) Yishay Mansour (Mansour@tau.ac.il) Teaching Assistance: Regev Schweiger

More information

Bayesian Decision Theory

Bayesian Decision Theory Bayesian Decision Theory 1/27 lecturer: authors: Jiri Matas, matas@cmp.felk.cvut.cz Václav Hlaváč, Jiri Matas Czech Technical University, Faculty of Electrical Engineering Department of Cybernetics, Center

More information

Favoring Consensus and Penalizing Disagreement in Group Decision Making

Favoring Consensus and Penalizing Disagreement in Group Decision Making Favoring Consensus and Penalizing Disagreement in Group Decision Making Paper: jc*-**-**-**** Favoring Consensus and Penalizing Disagreement in Group Decision Making José Luis García-Lapresta PRESAD Research

More information

Counter-examples to Dempster s rule of combination

Counter-examples to Dempster s rule of combination Jean Dezert 1, Florentin Smarandache 2, Mohammad Khoshnevisan 3 1 ONERA, 29 Av. de la Division Leclerc 92320, Chatillon, France 2 Department of Mathematics University of New Mexico Gallup, NM 8730, U.S.A.

More information

Algorithm-Independent Learning Issues

Algorithm-Independent Learning Issues Algorithm-Independent Learning Issues Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2007 c 2007, Selim Aksoy Introduction We have seen many learning

More information

ADVANCED METHODS FOR PATTERN RECOGNITION

ADVANCED METHODS FOR PATTERN RECOGNITION Università degli Studi di Cagliari Dottorato di Ricerca in Ingegneria Elettronica e Informatica XIV Ciclo ADVANCED METHODS FOR PATTERN RECOGNITION WITH REJECT OPTION Relatore Prof. Ing. Fabio ROLI Tesi

More information

A class of fusion rules based on the belief redistribution to subsets or complements

A class of fusion rules based on the belief redistribution to subsets or complements Chapter 5 A class of fusion rules based on the belief redistribution to subsets or complements Florentin Smarandache Chair of Math. & Sciences Dept., Univ. of New Mexico, 200 College Road, Gallup, NM 87301,

More information

A Simple Algorithm for Learning Stable Machines

A Simple Algorithm for Learning Stable Machines A Simple Algorithm for Learning Stable Machines Savina Andonova and Andre Elisseeff and Theodoros Evgeniou and Massimiliano ontil Abstract. We present an algorithm for learning stable machines which is

More information

Using Fuzzy Logic as a Complement to Probabilistic Radioactive Waste Disposal Facilities Safety Assessment -8450

Using Fuzzy Logic as a Complement to Probabilistic Radioactive Waste Disposal Facilities Safety Assessment -8450 Using Fuzzy Logic as a Complement to Probabilistic Radioactive Waste Disposal Facilities Safety Assessment -8450 F. L. De Lemos CNEN- National Nuclear Energy Commission; Rua Prof. Mario Werneck, s/n, BH

More information

Intelligent Systems Statistical Machine Learning

Intelligent Systems Statistical Machine Learning Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2014/2015, Our tasks (recap) The model: two variables are usually present: - the first one is typically discrete k

More information

MULTINOMIAL AGENT S TRUST MODELING USING ENTROPY OF THE DIRICHLET DISTRIBUTION

MULTINOMIAL AGENT S TRUST MODELING USING ENTROPY OF THE DIRICHLET DISTRIBUTION MULTINOMIAL AGENT S TRUST MODELING USING ENTROPY OF THE DIRICHLET DISTRIBUTION Mohammad Anisi 1 and Morteza Analoui 2 1 School of Computer Engineering, Iran University of Science and Technology, Narmak,

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,

More information

How to evaluate credit scorecards - and why using the Gini coefficient has cost you money

How to evaluate credit scorecards - and why using the Gini coefficient has cost you money How to evaluate credit scorecards - and why using the Gini coefficient has cost you money David J. Hand Imperial College London Quantitative Financial Risk Management Centre August 2009 QFRMC - Imperial

More information

A soft computing logic method for agricultural land suitability evaluation

A soft computing logic method for agricultural land suitability evaluation A soft computing logic method for agricultural land suitability evaluation B. Montgomery 1, S. Dragićević 1* and J. Dujmović 2 1 Geography Department, Simon Fraser University, 8888 University Drive, Burnaby,

More information

12. Vagueness, Uncertainty and Degrees of Belief

12. Vagueness, Uncertainty and Degrees of Belief 12. Vagueness, Uncertainty and Degrees of Belief KR & R Brachman & Levesque 2005 202 Noncategorical statements Ordinary commonsense knowledge quickly moves away from categorical statements like a P is

More information

Online Estimation of Discrete Densities using Classifier Chains

Online Estimation of Discrete Densities using Classifier Chains Online Estimation of Discrete Densities using Classifier Chains Michael Geilke 1 and Eibe Frank 2 and Stefan Kramer 1 1 Johannes Gutenberg-Universtität Mainz, Germany {geilke,kramer}@informatik.uni-mainz.de

More information

A Class of DSm Conditioning Rules 1

A Class of DSm Conditioning Rules 1 Class of DSm Conditioning Rules 1 Florentin Smarandache, Mark lford ir Force Research Laboratory, RIE, 525 Brooks Rd., Rome, NY 13441-4505, US bstract: In this paper we introduce two new DSm fusion conditioning

More information

Mapping Vulnerability and Risk of Mangrove Conversion to Pond Aquaculture in Myanmar

Mapping Vulnerability and Risk of Mangrove Conversion to Pond Aquaculture in Myanmar Aquaculture and Coastal Habitats Report No. 4 Mapping Vulnerability and Risk of Mangrove Conversion to Pond Aquaculture in Myanmar J. Ronald Eastman, Stefano Crema, Katherine Landesman Clark Labs, Clark

More information

Day 5: Generative models, structured classification

Day 5: Generative models, structured classification Day 5: Generative models, structured classification Introduction to Machine Learning Summer School June 18, 2018 - June 29, 2018, Chicago Instructor: Suriya Gunasekar, TTI Chicago 22 June 2018 Linear regression

More information

Final Examination CS540-2: Introduction to Artificial Intelligence

Final Examination CS540-2: Introduction to Artificial Intelligence Final Examination CS540-2: Introduction to Artificial Intelligence May 9, 2018 LAST NAME: SOLUTIONS FIRST NAME: Directions 1. This exam contains 33 questions worth a total of 100 points 2. Fill in your

More information

UNIVERSITY OF SURREY

UNIVERSITY OF SURREY UNIVERSITY OF SURREY B.Sc. Undergraduate Programmes in Computing B.Sc. Undergraduate Programmes in Mathematical Studies Level HE3 Examination MODULE CS364 Artificial Intelligence Time allowed: 2 hours

More information

Model Accuracy Measures

Model Accuracy Measures Model Accuracy Measures Master in Bioinformatics UPF 2017-2018 Eduardo Eyras Computational Genomics Pompeu Fabra University - ICREA Barcelona, Spain Variables What we can measure (attributes) Hypotheses

More information

Probabilistic Graphical Models for Image Analysis - Lecture 1

Probabilistic Graphical Models for Image Analysis - Lecture 1 Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.

More information

Evaluation. Andrea Passerini Machine Learning. Evaluation

Evaluation. Andrea Passerini Machine Learning. Evaluation Andrea Passerini passerini@disi.unitn.it Machine Learning Basic concepts requires to define performance measures to be optimized Performance of learning algorithms cannot be evaluated on entire domain

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

CSCI-567: Machine Learning (Spring 2019)

CSCI-567: Machine Learning (Spring 2019) CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Mar. 19, 2019 March 19, 2019 1 / 43 Administration March 19, 2019 2 / 43 Administration TA3 is due this week March

More information

Analysis of the Performance of AdaBoost.M2 for the Simulated Digit-Recognition-Example

Analysis of the Performance of AdaBoost.M2 for the Simulated Digit-Recognition-Example Analysis of the Performance of AdaBoost.M2 for the Simulated Digit-Recognition-Example Günther Eibl and Karl Peter Pfeiffer Institute of Biostatistics, Innsbruck, Austria guenther.eibl@uibk.ac.at Abstract.

More information

Evidence with Uncertain Likelihoods

Evidence with Uncertain Likelihoods Evidence with Uncertain Likelihoods Joseph Y. Halpern Cornell University Ithaca, NY 14853 USA halpern@cs.cornell.edu Riccardo Pucella Cornell University Ithaca, NY 14853 USA riccardo@cs.cornell.edu Abstract

More information

Bayesian Reasoning and Recognition

Bayesian Reasoning and Recognition Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIAG 2 / osig 1 Second Semester 2013/2014 Lesson 12 28 arch 2014 Bayesian Reasoning and Recognition Notation...2 Pattern Recognition...3

More information

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 1 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification

More information

Ensembles of Classifiers.

Ensembles of Classifiers. Ensembles of Classifiers www.biostat.wisc.edu/~dpage/cs760/ 1 Goals for the lecture you should understand the following concepts ensemble bootstrap sample bagging boosting random forests error correcting

More information

A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE

A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE Li Sheng Institute of intelligent information engineering Zheiang University Hangzhou, 3007, P. R. China ABSTRACT In this paper, a neural network-driven

More information

Machine Learning and Data Mining. Linear classification. Kalev Kask

Machine Learning and Data Mining. Linear classification. Kalev Kask Machine Learning and Data Mining Linear classification Kalev Kask Supervised learning Notation Features x Targets y Predictions ŷ = f(x ; q) Parameters q Program ( Learner ) Learning algorithm Change q

More information

Published in: Tenth Tbilisi Symposium on Language, Logic and Computation: Gudauri, Georgia, September 2013

Published in: Tenth Tbilisi Symposium on Language, Logic and Computation: Gudauri, Georgia, September 2013 UvA-DARE (Digital Academic Repository) Estimating the Impact of Variables in Bayesian Belief Networks van Gosliga, S.P.; Groen, F.C.A. Published in: Tenth Tbilisi Symposium on Language, Logic and Computation:

More information

Small sample size generalization

Small sample size generalization 9th Scandinavian Conference on Image Analysis, June 6-9, 1995, Uppsala, Sweden, Preprint Small sample size generalization Robert P.W. Duin Pattern Recognition Group, Faculty of Applied Physics Delft University

More information

The maximum Deng entropy

The maximum Deng entropy The maximum Deng entropy Bingyi Kang a, Yong Deng a,b,c, a School of Computer and Information Science, Southwest University, Chongqing, 40075, China b School of Electronics and Information, Northwestern

More information