Minimal Attribute Space Bias for Attribute Reduction

Size: px
Start display at page:

Download "Minimal Attribute Space Bias for Attribute Reduction"

Transcription

1 Minimal Attribute Space Bias for Attribute Reduction Fan Min, Xianghui Du, Hang Qiu, and Qihe Liu School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu , China {minfan, xianghd, qiuhang, Abstract. Attribute reduction is an important inductive learning issue addressed by the Rough Sets society. Most existing works on this issue use the minimal attribute bias, i.e., searching for reducts with the minimal number of attributes. But this bias does not work well for datasets where different attributes have different sizes of domains. In this paper, we propose a more reasonable strategy called the minimal attribute space bias, i.e., searching for reducts with the minimal attribute domain sizes product. In most cases, this bias can help to obtain reduced decision tables with the best space coverage, thus helpful for obtaining small rule sets with good predicting performance. Empirical study on some datasets validates our analysis. Keywords. Attribute reduction, bias, space coverage, rule set. 1 Introduction Practical machine learning algorithms are known to degrade in performance (prediction accuracy) when faced with many attributes that are not necessary for rule discovery [1]. It is therefore not surprising that much research has been carried out on attribute reduction [2], particularly by people in the Rough Sets society. A reduct is a subset of attributes that is jointly sufficient and individually necessary for preserving the same information (in terms of positive region [3], class distribution [4] among others) under consideration as provided by the entire set of attributes. [5] A commonly used reduct selection/construction strategy, called the minimal attribute bias [1][3][6], is to searching for a reduct with the minimal number of attributes. In some cases, especially when different attribute have approximately the same size of domain, this bias may be helpful for obtaining small rule sets with good performance. However, for data in reality where attribute domain sizes vary, this strategy is unfair since attributes with larger domains tend to have better discernibility or other significance measures [7], and it has severe implications when applied blindly without regarding for the resulting induced concept [1]. To cope with these problems, in this paper we propose a new bias called the minimal attribute space bias which is intended to minimize the attribute space. We argue that this bias is more reasonable, thus more helpful for obtaining small

2 rule set, than the minimal attribute bias. Empirical study on some datasets in the UCI library [8] validates our analysis. 2 Preliminaries In this section we enumerate basic concepts introduced by Pawlak [3] through an example. Formally, a decision table is a triple S = (U, C, {d}) where d C is the decision attribute and elements of C are called conditional attributes or simply conditions. Table 1 lists a decision table where U = {t1,..., t8}, C = {Shape, Material, Weight, Color} and d = Size. Table 1. An exemplary decision table Toy Shape Material Weight Color Size t1 round wood light red small t2 round plastic heavy black small t3 round wood heavy white large t4 round wood light white small t5 triangle wood light green small t6 triangle plastic heavy blue large t7 triangle plastic light pink large t8 triangle plastic heavy yellow large Any B C {d} determines a binary relation I(B) on U, which will be called an indiscernibility relation, and is defined as follows: I(B) = {(x i, x j ) U U a B, a(x i ) = a(x j )}, (1) where a(x) denotes the value of attribute a for element x. A partition determined by B is denoted by U/I(B), or simply by U/B. Let BX denotes B lower approximation of X, the positive region of {d} with respect to B C is defined as P OS B ({d}) = X U/{d} B(X). A reduct is the minimal subset of attributes that enables the same classification of elements of the universe as the whole set of attributes. This can be formally defined as follows: Definition 1. Any B C is called a decision relative reduct of S = (U, C, {d}) iff: 1. P OS B ({d}) = P OS C ({d}), and 2. a B, P OS B {a} ({d}) P OS C ({d}). A decision relative reduct can be simply called a relative reduct, or a reduct for briefness if the decision attribute is obvious from the context. According to Definition 1, the exemplary decision table has two reducts: R 1 ={Shape, Material, Weight} and R 2 ={Weight, Color}.

3 3 The Minimal Attribute Bias This bias is described as follows: A reduct R is optimal iff R is minimal, where denotes the cardinality of a set. According to this bias, R 2 ={Weight, Color} is an optimal reduct of Table 1 because R 1 = 3 and R 2 = 2. For the sake of clarity, we use the term minimal reduct instead of optimal reduct in the following context. The main advantage of this bias is simple and tend to give short rules. For datasets where different attributes have approximately the same size of domains, it may be also helpful for obtaining small rule sets with good predicting performance. However, it also has the following drawbacks: 1. Unfair for different attributes. For example, in Table 1, attribute Color is the most important attribute from the viewpoint of discernibility. But this is due to its relatively large domain (7 values versus 2 of others) rather than its intrinsic importance. 2. Too many optimal reducts. For example, the Mushroom dataset [8] (further discussed in Subsection 5.1) has 14 minimal reducts. Some of them perform well in terms of further rule set generation and/or decision tree construction, but others do not. The bias does not indicate a more detailed strategy for choosing among those reducts. 4 The Minimal Attribute Space Bias We propose the minimal attribute space bias as follows: A reduct R is optimal iff Π a R V a is minimal, where V a is the domain of attribute a. According to this bias, R 1 = {Shape, Material, Weight} is an optimal reduct of Table 1 because Π a R1 V a = 8 and Π a R2 V a = 14. We also use the term minimal space reduct instead of optimal reduct. Remark 1. If V ai = V aj for any a i, a j C, then the minimal attribute space bias coincides with the minimal attribute bias. Now we explain why this bias is more reasonable than the minimal attribute bias using the exemplary decision table. Each object in the table corresponds with a decision rule. For example, t1 corresponds with Shape = round Material = wood Weight = light Color = red Size = small. This type of rules will be called original rules since no inductive learning algorithm has been introduced. Because no object pairs are indiscernible, there are a total of 8 original rules. On the other hand, the attribute space of the decision table is Shape Material Weight Color = = 56. Therefore objects in the decision table only cover 8/56 = 1/7 = of the attribute space, and the original rule set may have poor performance in terms of coverage.

4 As an inductive approach, attribute reduction can reduce the number of attribute; or more importantly, it can reduce the attribute space. The attribute space of S(R 1 ) = {{t1,..., t8}, R 1, {Size}} is Shape Material Weight = = 8, while it has two indiscernible object pairs: t1 and t4 as well as t6 and t8, hence only 8 2 = 6 original rules could be obtained, incurring 6/8 = 0.75 of space coverage. The attribute space of S(R 2 ) = {{t1,..., t8}, R 2, {Size}} is Weight Color = 2 7 = 14, and no indiscernible object pairs exist, hence 8 original rules could be obtained, incurring 8/14 = of space coverage. From this viewpoint, both reducts have notable generalization ability, and R 1 performs better (with space coverage 0.75 versus of R 2 ). Table 2. Rule sets generated from S(R 1) and S(R 2) rule No. rule support r1 Material = wood Weight = light Size = small 3 r2 Shape = triangle Material = plastic Size = large 3 r3 Shape = round Weight = light Size = small 2 r4 Shape = triangle Weight = heavy Size = large 2 r5 Shape = round Material = plastic Size = small 1 r6 Material = wood Weight = heavy Size = large 1 r7 Shape = triangle Material = wood Size = small 1 r8 material = plastic weight = light Size = large 1 r9 Color = red Size = small 1 r10 Color = black Size = small 1 r11 Weight = heavy Color = white Size = large 1 r12 Weight = light Color = white Size = small 1 r13 Color = green Size = small 1 r14 Color = blue Size = large 1 r15 Color = pink Size = large 1 r16 Color = yellow Size = large 1 It should be noted that better generalization ability does not ensure smaller rule sets. In fact, using the exhaustive algorithm [9] we obtained 8 rules for either reduced decision tables, as listed in Table 2, where the former 8 rules corresponds with S(R 1 ). But it should be noted further that each rule generated from S(R 2 ) is supported by only 1 object, while in all rules generated from S(R 1 ), 2 rules (r1 and r2) are supported by 3 objects, and another 2 rules (r3 and r4) are supported by 2 objects. In other words, R 1 is more helpful for generating strong rules. One can observe that in this case both rule sets cover the whole attribute space. While for larger datasets, reducts with smaller attribute spaces often result in smaller rule sets with better space coverage. Formally, the space coverage of S = (U, C, {d}) is defined as SC(S) = U/C Π a C V a, (2)

5 which serves as an important factor for further rule generation / decision tree construction. From the viewpoint of space coverage, the goal of attribute reduction should be searching for a reduct R such that SC((U, R, {d})) is minimal. One approach is to maximize U/R, but U/R has an upper bound U, and U/R does not vary too much for different reducts. Thus this approach does not make sense. The other approach is to minimize Π a R V a, which coincides with the minimal attribute space bias. In most cases, minimal attribute space results in maximal space coverage. One might construct a counterexample as follows: A decision table S = (U, C, {d}) and two reducts R 1, R 2 satisfying U/R 1 > U/R 2, Π a R1 V a > Π a R2 V a and SC((U, R 1, {d})) > SC((U, R 2, {d})); but this situation is quite unlikely to happen in real data. 5 Experiments and Comparisons There are very few datasets (e.g., Nursery) with complete coverage of the attribute space. For many datasets, attribute reduction is a very important approach to improving the space coverage. We tested these two reduct selection biases on some datasets of the UCI library [8] using RSES [9] and a software developed by us called RDK (Rough Developer s Kit). For some datasets the set of all minimal space reducts coincides with the set of all minimal reducts. These datasets can be classified as follows: 1. One reduct datasets, e.g., Zoo, solar-flare and Monks; 2. Datasets whose all conditional attributes have the same size of domain, e.g., Letter Recognition and Tic-Tac-Toe, and 3. Other Datasets, e.g., bridges. In what follows experiments on two datasets will be discussed in more detail. 5.1 Experiments on Mushroom The Mushroom dataset [8] contains 8416 objects and 22 conditional attributes. The domain sizes of attributes vary from 1 (veil-type, the attribute value UNIVERSAL announced in agaricus-lepiota.names never appeared in the dataset) to 12 (gill-color). It has 292 reducts, and all minimal space reducts are also minimal reducts. For each minimal reduct, LEM2 was employed (with the cover parameter set to 1.0) to generate rule sets on reduced decision tables. Furthermore, we used CV-5 (the rule generation algorithm was also LEM2) to test the performance of those reducts. For all reducts tested, both the coverage and the accuracy of respective rule sets were 1.0. Other results are listed in Table 3. For this dataset, the number of minimal attribute reducts is much less than the number of minimal reducts. Also, minimal attribute reducts are more helpful for obtaining smaller rule sets.

6 Table 3. Experimental results of Mushroom minimal attribute bias minimal attribute space bias optimal reducts 14 2 minimal rule set size maximal rule set size average rule set size Experiments on Soybean The Large Soybean Database [8] contains two parts: the training set with 307 instances, and the testing set with 376 instances. There are 35 nominal conditional attributes, with domain sizes varying from 2 to 8, and some of them have missing (unknown) values. The domain size (i.e., 18) of the decision attribute is rather large. Due to the relatively large number of attributes, when we tried to use the exhaustive algorithm [9] to obtain the set of all reducts, an error out of memory was reported. So we used the genetic algorithm on the training set to obtain reducts, with the speed set to low and the number of reducts set to 100. In this way, 100 reducts were obtained, 35 of which were minimal reducts, and 9 of which were minimal space reducts. 8 out of 9 minimal space reducts contain 9 attributes, hence they were also minimal reducts. Rule sets were generated through employing the exhaustive algorithm on reduced decision tables, then they were tested on the testing set. Some results are listed in Table 4. Table 4. Experimental results of Soybean (Bolded Values Indicate the Best Results) minimal attribute reducts minimal attribute space reducts minimal maximal average minimal maximal average rule set size coverage accuracy F -measure Since the minimal attribute reduct set is much larger than the minimal attribute space reduct set, it contains both the best and the worst reducts. In general, for this dataset the minimal attribute space bias outperforms the minimal attribute bias in terms of average rule set size (325 less), average accuracy (0.024 more) and averge F -measure (0.024 more). It is quite interesting that the latter bias outperforms (0.004 more than) the former in terms of average coverage. Two reducts drew our special attention:

7 1. The best reduct. It helped obtaining a rule set with a predication accuracy of and F -measure of 0.740, and it was included in both reduct sets; and 2. The minimal attribute space reduct with 10 attributes. Although not a minimal reduct, it helped obtain a rule set containing 1789 rules, with the predication coverage 0.949, accuracy and F -measure The results are fairly good compared with that of minimal reducts. 6 Discussion In this section we discuss these two biases from a broader viewpoint. Discretization scheme Reduct Rule set Predicting performance Decision tree Fig. 1. A typical inductive learning scenario As depicted in Fig. 1, the ultimate goal of inductive learning is to obtain good predicting performance, defined by the coverage, the accuracy, or the combination of both (e.g., F -measure) on the data. But the predicting performance can be obtained only after rule set was generated or decision tree [10] was constructed (for the sake of simplicity, other approaches such as neural network or knn are not included in Fig. 1). According to Occam s Razor, smaller rule sets, or simpler decision trees (with least nodes) are preferred. In order to obtain a small rule set or a simple decision tree, also according to Occam s Razor, the simplest reduct is desired. But the key issue is: What is the metric of evaluating the simplicity of a reduct? Aforementioned biases are two metrics, among which the new bias seems to be closer to the essence. Then why the minimal attribute bias worked well for so many applications? In fact, many reduct construction algorithms use the following strategy: [i]f two attributes have the same performance with respect to the criterion described above then the one having less values is selected [1], and it is quite possible that a minimal space reduct be constructed while a minimal reduct is required. Moreover, even if the minimal reduct obtained is not a minimal space reduct, its attribute spaces is not too large compared with that of a minimal space reduct. In other words, the minimal attribute bias is often a good approximation of the minimal attribute space bias.

8 7 Conclusions and Further Works Compared with the minimal attribute bias, the minimal attribute space bias is closer to goal of constructing simple reducts from viewpoints of attribute space and attribute space coverage. Also, it does not incur the fairness problem. Experiments on two datasets showed that the new bias can help to narrow the scope of optimal reducts, and more importantly, it can help to obtain better rule sets in terms of accuracy and F -measure. Since the definition of a bias is a quite fundamental issue, many research works, e.g., reduct constructing algorithms, on the new bias are expected in the near future. Acknowledgement Fan Min was supported by an information distribution project under grant No. 9140A DZ223 and the Youth Foundation of UESTC. The authors would like to thank Zichun Zhong and Yue Liu for their help in experiments and paper proofing. References 1. Zhong, N., Dong, J.: Using rough sets with heuristics for feature selection. Journal of Intelligent Information Systems, 16 (2001) Pawlak, Z.: Rough sets. International Journal of Computer and Information Sciences 11 (1982) Pawlak, Z.: Some issues on rough sets. In Peters, J.F., Skowron, A., Grzyma la- Busse, J.W., Kostek, B., Świniarski, R.W., Szczuka, M.S., eds.: Transactions on Rough Sets I. LNCS Springer-Verlag, Berlin Heidelberg (2004) Zhang, W., Mi, J., Wu, W.: Knowledge reductions in inconsistent information systems. Chinese Journal of Computers 26(1) (2003) Yao, Y., Yan, Z., Wang, J.: On reduct construction algorithms. In Wang, G., Peters, J.F., Skowron, A., Yao, Y., eds.: RSKT LNCS 4062, Berlin Heidelberg, Springer-Verlag (2006) Skowron, A., Rauszer, C.: The discernibility matrices and functions in information systems. In S lowiński, R., ed.: Intelligent Decision Support Handbook of Applications and Advances of the Rough Sets Theory. Kluwer, Academic Publishers, Dordrecht (1992) Xu, C., Min, F.: Weighted reduction for decision tables. In Wang, L., Jiao, L., Shi, G., Li, X., Liu, J., eds.: FSKD LNCS 4223, Berlin Heidelberg, Springer-Verlag (2006) Blake, C.L., Merz, C.J.: UCI repository of machine learning databases, mlearn/mlrepository.html (1998) 9. Bazan, J., Szczuka, M.: The RSES homepage, rses ( ) 10. Quinlan, J.R.: Induction of decision trees. Machine Learning 1 (1986)

Easy Categorization of Attributes in Decision Tables Based on Basic Binary Discernibility Matrix

Easy Categorization of Attributes in Decision Tables Based on Basic Binary Discernibility Matrix Easy Categorization of Attributes in Decision Tables Based on Basic Binary Discernibility Matrix Manuel S. Lazo-Cortés 1, José Francisco Martínez-Trinidad 1, Jesús Ariel Carrasco-Ochoa 1, and Guillermo

More information

Research on Complete Algorithms for Minimal Attribute Reduction

Research on Complete Algorithms for Minimal Attribute Reduction Research on Complete Algorithms for Minimal Attribute Reduction Jie Zhou, Duoqian Miao, Qinrong Feng, and Lijun Sun Department of Computer Science and Technology, Tongji University Shanghai, P.R. China,

More information

Interpreting Low and High Order Rules: A Granular Computing Approach

Interpreting Low and High Order Rules: A Granular Computing Approach Interpreting Low and High Order Rules: A Granular Computing Approach Yiyu Yao, Bing Zhou and Yaohua Chen Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 E-mail:

More information

Classification Based on Logical Concept Analysis

Classification Based on Logical Concept Analysis Classification Based on Logical Concept Analysis Yan Zhao and Yiyu Yao Department of Computer Science, University of Regina, Regina, Saskatchewan, Canada S4S 0A2 E-mail: {yanzhao, yyao}@cs.uregina.ca Abstract.

More information

On Improving the k-means Algorithm to Classify Unclassified Patterns

On Improving the k-means Algorithm to Classify Unclassified Patterns On Improving the k-means Algorithm to Classify Unclassified Patterns Mohamed M. Rizk 1, Safar Mohamed Safar Alghamdi 2 1 Mathematics & Statistics Department, Faculty of Science, Taif University, Taif,

More information

CRITERIA REDUCTION OF SET-VALUED ORDERED DECISION SYSTEM BASED ON APPROXIMATION QUALITY

CRITERIA REDUCTION OF SET-VALUED ORDERED DECISION SYSTEM BASED ON APPROXIMATION QUALITY International Journal of Innovative omputing, Information and ontrol II International c 2013 ISSN 1349-4198 Volume 9, Number 6, June 2013 pp. 2393 2404 RITERI REDUTION OF SET-VLUED ORDERED DEISION SYSTEM

More information

Feature Selection with Fuzzy Decision Reducts

Feature Selection with Fuzzy Decision Reducts Feature Selection with Fuzzy Decision Reducts Chris Cornelis 1, Germán Hurtado Martín 1,2, Richard Jensen 3, and Dominik Ślȩzak4 1 Dept. of Mathematics and Computer Science, Ghent University, Gent, Belgium

More information

Iterative Laplacian Score for Feature Selection

Iterative Laplacian Score for Feature Selection Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,

More information

International Journal of Approximate Reasoning

International Journal of Approximate Reasoning International Journal of Approximate Reasoning 49 (2008) 689 700 Contents lists available at ScienceDirect International Journal of Approximate Reasoning journal homepage: www.elsevier.com/locate/ijar

More information

Ensembles of classifiers based on approximate reducts

Ensembles of classifiers based on approximate reducts Fundamenta Informaticae 34 (2014) 1 10 1 IOS Press Ensembles of classifiers based on approximate reducts Jakub Wróblewski Polish-Japanese Institute of Information Technology and Institute of Mathematics,

More information

Classification of Voice Signals through Mining Unique Episodes in Temporal Information Systems: A Rough Set Approach

Classification of Voice Signals through Mining Unique Episodes in Temporal Information Systems: A Rough Set Approach Classification of Voice Signals through Mining Unique Episodes in Temporal Information Systems: A Rough Set Approach Krzysztof Pancerz, Wies law Paja, Mariusz Wrzesień, and Jan Warcho l 1 University of

More information

METRIC BASED ATTRIBUTE REDUCTION IN DYNAMIC DECISION TABLES

METRIC BASED ATTRIBUTE REDUCTION IN DYNAMIC DECISION TABLES Annales Univ. Sci. Budapest., Sect. Comp. 42 2014 157 172 METRIC BASED ATTRIBUTE REDUCTION IN DYNAMIC DECISION TABLES János Demetrovics Budapest, Hungary Vu Duc Thi Ha Noi, Viet Nam Nguyen Long Giang Ha

More information

Selected Algorithms of Machine Learning from Examples

Selected Algorithms of Machine Learning from Examples Fundamenta Informaticae 18 (1993), 193 207 Selected Algorithms of Machine Learning from Examples Jerzy W. GRZYMALA-BUSSE Department of Computer Science, University of Kansas Lawrence, KS 66045, U. S. A.

More information

2 WANG Jue, CUI Jia et al. Vol.16 no", the discernibility matrix is only a new kind of learning method. Otherwise, we have to provide the specificatio

2 WANG Jue, CUI Jia et al. Vol.16 no, the discernibility matrix is only a new kind of learning method. Otherwise, we have to provide the specificatio Vol.16 No.1 J. Comput. Sci. & Technol. Jan. 2001 Investigation on AQ11, ID3 and the Principle of Discernibility Matrix WANG Jue (Ξ ±), CUI Jia ( ) and ZHAO Kai (Π Λ) Institute of Automation, The Chinese

More information

ARPN Journal of Science and Technology All rights reserved.

ARPN Journal of Science and Technology All rights reserved. Rule Induction Based On Boundary Region Partition Reduction with Stards Comparisons Du Weifeng Min Xiao School of Mathematics Physics Information Engineering Jiaxing University Jiaxing 34 China ABSTRACT

More information

On Granular Rough Computing: Factoring Classifiers through Granulated Decision Systems

On Granular Rough Computing: Factoring Classifiers through Granulated Decision Systems On Granular Rough Computing: Factoring Classifiers through Granulated Decision Systems Lech Polkowski 1,2, Piotr Artiemjew 2 Department of Mathematics and Computer Science University of Warmia and Mazury

More information

Occam s Razor Just Got Sharper

Occam s Razor Just Got Sharper Occam s Razor Just Got Sharper Saher Esmeir and Shaul Markovitch Computer Science Department, Technion Israel Institute of Technology, Haifa 32000, Israel {esaher, shaulm}@cs.technion.ac.il Abstract Occam

More information

ENTROPIES OF FUZZY INDISCERNIBILITY RELATION AND ITS OPERATIONS

ENTROPIES OF FUZZY INDISCERNIBILITY RELATION AND ITS OPERATIONS International Journal of Uncertainty Fuzziness and Knowledge-Based Systems World Scientific ublishing Company ENTOIES OF FUZZY INDISCENIBILITY ELATION AND ITS OEATIONS QINGUA U and DAEN YU arbin Institute

More information

Aijun An and Nick Cercone. Department of Computer Science, University of Waterloo. methods in a context of learning classication rules.

Aijun An and Nick Cercone. Department of Computer Science, University of Waterloo. methods in a context of learning classication rules. Discretization of Continuous Attributes for Learning Classication Rules Aijun An and Nick Cercone Department of Computer Science, University of Waterloo Waterloo, Ontario N2L 3G1 Canada Abstract. We present

More information

A new Approach to Drawing Conclusions from Data A Rough Set Perspective

A new Approach to Drawing Conclusions from Data A Rough Set Perspective Motto: Let the data speak for themselves R.A. Fisher A new Approach to Drawing Conclusions from Data A Rough et Perspective Zdzisław Pawlak Institute for Theoretical and Applied Informatics Polish Academy

More information

Naive Bayesian Rough Sets

Naive Bayesian Rough Sets Naive Bayesian Rough Sets Yiyu Yao and Bing Zhou Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 {yyao,zhou200b}@cs.uregina.ca Abstract. A naive Bayesian classifier

More information

Finite Mixture Model of Bounded Semi-naive Bayesian Networks Classifier

Finite Mixture Model of Bounded Semi-naive Bayesian Networks Classifier Finite Mixture Model of Bounded Semi-naive Bayesian Networks Classifier Kaizhu Huang, Irwin King, and Michael R. Lyu Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,

More information

Granular Computing: Granular Classifiers and Missing Values

Granular Computing: Granular Classifiers and Missing Values 1 Granular Computing: Granular Classifiers and Missing Values Lech Polkowski 1,2 and Piotr Artiemjew 2 Polish-Japanese Institute of Information Technology 1 Koszykowa str. 86, 02008 Warsaw, Poland; Department

More information

Machine Learning & Data Mining

Machine Learning & Data Mining Group M L D Machine Learning M & Data Mining Chapter 7 Decision Trees Xin-Shun Xu @ SDU School of Computer Science and Technology, Shandong University Top 10 Algorithm in DM #1: C4.5 #2: K-Means #3: SVM

More information

An Empirical Study of Building Compact Ensembles

An Empirical Study of Building Compact Ensembles An Empirical Study of Building Compact Ensembles Huan Liu, Amit Mandvikar, and Jigar Mody Computer Science & Engineering Arizona State University Tempe, AZ 85281 {huan.liu,amitm,jigar.mody}@asu.edu Abstract.

More information

Parts 3-6 are EXAMPLES for cse634

Parts 3-6 are EXAMPLES for cse634 1 Parts 3-6 are EXAMPLES for cse634 FINAL TEST CSE 352 ARTIFICIAL INTELLIGENCE Fall 2008 There are 6 pages in this exam. Please make sure you have all of them INTRODUCTION Philosophical AI Questions Q1.

More information

A novel k-nn approach for data with uncertain attribute values

A novel k-nn approach for data with uncertain attribute values A novel -NN approach for data with uncertain attribute values Asma Trabelsi 1,2, Zied Elouedi 1, and Eric Lefevre 2 1 Université de Tunis, Institut Supérieur de Gestion de Tunis, LARODEC, Tunisia trabelsyasma@gmail.com,zied.elouedi@gmx.fr

More information

IN the areas of machine learning, artificial intelligence, as

IN the areas of machine learning, artificial intelligence, as INTL JOURNAL OF ELECTRONICS AND TELECOMMUNICATIONS, 22, VOL. 58, NO., PP. 7 76 Manuscript received December 3, 2; revised March 22. DOI:.2478/v77-2--x Features Reduction Using Logic Minimization Techniques

More information

Computers and Mathematics with Applications

Computers and Mathematics with Applications Computers and Mathematics with Applications 59 (2010) 431 436 Contents lists available at ScienceDirect Computers and Mathematics with Applications journal homepage: www.elsevier.com/locate/camwa A short

More information

A Scientometrics Study of Rough Sets in Three Decades

A Scientometrics Study of Rough Sets in Three Decades A Scientometrics Study of Rough Sets in Three Decades JingTao Yao and Yan Zhang Department of Computer Science University of Regina [jtyao, zhang83y]@cs.uregina.ca Oct. 8, 2013 J. T. Yao & Y. Zhang A Scientometrics

More information

Three Discretization Methods for Rule Induction

Three Discretization Methods for Rule Induction Three Discretization Methods for Rule Induction Jerzy W. Grzymala-Busse, 1, Jerzy Stefanowski 2 1 Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, Kansas 66045

More information

Machine Learning on temporal data

Machine Learning on temporal data Machine Learning on temporal data Classification rees for ime Series Ahlame Douzal (Ahlame.Douzal@imag.fr) AMA, LIG, Université Joseph Fourier Master 2R - MOSIG (2011) Plan ime Series classification approaches

More information

An algorithm for induction of decision rules consistent with the dominance principle

An algorithm for induction of decision rules consistent with the dominance principle An algorithm for induction of decision rules consistent with the dominance principle Salvatore Greco 1, Benedetto Matarazzo 1, Roman Slowinski 2, Jerzy Stefanowski 2 1 Faculty of Economics, University

More information

Transversal and Function Matroidal Structures of Covering-Based Rough Sets

Transversal and Function Matroidal Structures of Covering-Based Rough Sets Transversal and Function Matroidal Structures of Covering-Based Rough Sets Shiping Wang 1, William Zhu 2,,andFanMin 2 1 School of Mathematical Sciences, University of Electronic Science and Technology

More information

ROUGH set methodology has been witnessed great success

ROUGH set methodology has been witnessed great success IEEE TRANSACTIONS ON FUZZY SYSTEMS, VOL. 14, NO. 2, APRIL 2006 191 Fuzzy Probabilistic Approximation Spaces and Their Information Measures Qinghua Hu, Daren Yu, Zongxia Xie, and Jinfu Liu Abstract Rough

More information

Study on Classification Methods Based on Three Different Learning Criteria. Jae Kyu Suhr

Study on Classification Methods Based on Three Different Learning Criteria. Jae Kyu Suhr Study on Classification Methods Based on Three Different Learning Criteria Jae Kyu Suhr Contents Introduction Three learning criteria LSE, TER, AUC Methods based on three learning criteria LSE:, ELM TER:

More information

Randomized Decision Trees

Randomized Decision Trees Randomized Decision Trees compiled by Alvin Wan from Professor Jitendra Malik s lecture Discrete Variables First, let us consider some terminology. We have primarily been dealing with real-valued data,

More information

Home Page. Title Page. Page 1 of 35. Go Back. Full Screen. Close. Quit

Home Page. Title Page. Page 1 of 35. Go Back. Full Screen. Close. Quit JJ II J I Page 1 of 35 General Attribute Reduction of Formal Contexts Tong-Jun Li Zhejiang Ocean University, China litj@zjou.edu.cn September, 2011,University of Milano-Bicocca Page 2 of 35 Objective of

More information

The size of decision table can be understood in terms of both cardinality of A, denoted by card (A), and the number of equivalence classes of IND (A),

The size of decision table can be understood in terms of both cardinality of A, denoted by card (A), and the number of equivalence classes of IND (A), Attribute Set Decomposition of Decision Tables Dominik Slezak Warsaw University Banacha 2, 02-097 Warsaw Phone: +48 (22) 658-34-49 Fax: +48 (22) 658-34-48 Email: slezak@alfa.mimuw.edu.pl ABSTRACT: Approach

More information

EVALUATING MISCLASSIFICATION PROBABILITY USING EMPIRICAL RISK 1. Victor Nedel ko

EVALUATING MISCLASSIFICATION PROBABILITY USING EMPIRICAL RISK 1. Victor Nedel ko 94 International Journal "Information Theories & Applications" Vol13 [Raudys, 001] Raudys S, Statistical and neural classifiers, Springer, 001 [Mirenkova, 00] S V Mirenkova (edel ko) A method for prediction

More information

Quantization of Rough Set Based Attribute Reduction

Quantization of Rough Set Based Attribute Reduction A Journal of Software Engineering and Applications, 0, 5, 7 doi:46/sea05b0 Published Online Decemer 0 (http://wwwscirporg/ournal/sea) Quantization of Rough Set Based Reduction Bing Li *, Peng Tang, Tommy

More information

Multiple Similarities Based Kernel Subspace Learning for Image Classification

Multiple Similarities Based Kernel Subspace Learning for Image Classification Multiple Similarities Based Kernel Subspace Learning for Image Classification Wang Yan, Qingshan Liu, Hanqing Lu, and Songde Ma National Laboratory of Pattern Recognition, Institute of Automation, Chinese

More information

Bayesian Networks for Classification

Bayesian Networks for Classification Finite Mixture Model of Bounded Semi-Naive Bayesian Networks for Classification Kaizhu Huang, Irwin King and Michael R. Lyu Department of Computer Science and Engineering The Chinese University of Hong

More information

A Generalized Decision Logic in Interval-set-valued Information Tables

A Generalized Decision Logic in Interval-set-valued Information Tables A Generalized Decision Logic in Interval-set-valued Information Tables Y.Y. Yao 1 and Qing Liu 2 1 Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 E-mail: yyao@cs.uregina.ca

More information

Data mining using Rough Sets

Data mining using Rough Sets Data mining using Rough Sets Alber Sánchez 1 alber.ipia@inpe.br 1 Instituto Nacional de Pesquisas Espaciais, São José dos Campos, SP, Brazil Referata Geoinformatica, 2015 1 / 44 Table of Contents Rough

More information

Similarity-based Classification with Dominance-based Decision Rules

Similarity-based Classification with Dominance-based Decision Rules Similarity-based Classification with Dominance-based Decision Rules Marcin Szeląg, Salvatore Greco 2,3, Roman Słowiński,4 Institute of Computing Science, Poznań University of Technology, 60-965 Poznań,

More information

Notes on Rough Set Approximations and Associated Measures

Notes on Rough Set Approximations and Associated Measures Notes on Rough Set Approximations and Associated Measures Yiyu Yao Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 E-mail: yyao@cs.uregina.ca URL: http://www.cs.uregina.ca/

More information

Boosting: Foundations and Algorithms. Rob Schapire

Boosting: Foundations and Algorithms. Rob Schapire Boosting: Foundations and Algorithms Rob Schapire Example: Spam Filtering problem: filter out spam (junk email) gather large collection of examples of spam and non-spam: From: yoav@ucsd.edu Rob, can you

More information

Foundations of Classification

Foundations of Classification Foundations of Classification J. T. Yao Y. Y. Yao and Y. Zhao Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 {jtyao, yyao, yanzhao}@cs.uregina.ca Summary. Classification

More information

LBR-Meta: An Efficient Algorithm for Lazy Bayesian Rules

LBR-Meta: An Efficient Algorithm for Lazy Bayesian Rules LBR-Meta: An Efficient Algorithm for Lazy Bayesian Rules Zhipeng Xie School of Computer Science Fudan University 220 Handan Road, Shanghai 200433, PR. China xiezp@fudan.edu.cn Abstract LBR is a highly

More information

High Frequency Rough Set Model based on Database Systems

High Frequency Rough Set Model based on Database Systems High Frequency Rough Set Model based on Database Systems Kartik Vaithyanathan kvaithya@gmail.com T.Y.Lin Department of Computer Science San Jose State University San Jose, CA 94403, USA tylin@cs.sjsu.edu

More information

Learning Decision Trees

Learning Decision Trees Learning Decision Trees Machine Learning Spring 2018 1 This lecture: Learning Decision Trees 1. Representation: What are decision trees? 2. Algorithm: Learning decision trees The ID3 algorithm: A greedy

More information

Decision Trees. Lewis Fishgold. (Material in these slides adapted from Ray Mooney's slides on Decision Trees)

Decision Trees. Lewis Fishgold. (Material in these slides adapted from Ray Mooney's slides on Decision Trees) Decision Trees Lewis Fishgold (Material in these slides adapted from Ray Mooney's slides on Decision Trees) Classification using Decision Trees Nodes test features, there is one branch for each value of

More information

ROUGH SET THEORY FOR INTELLIGENT INDUSTRIAL APPLICATIONS

ROUGH SET THEORY FOR INTELLIGENT INDUSTRIAL APPLICATIONS ROUGH SET THEORY FOR INTELLIGENT INDUSTRIAL APPLICATIONS Zdzisław Pawlak Institute of Theoretical and Applied Informatics, Polish Academy of Sciences, Poland, e-mail: zpw@ii.pw.edu.pl ABSTRACT Application

More information

ROUGH SETS THEORY AND DATA REDUCTION IN INFORMATION SYSTEMS AND DATA MINING

ROUGH SETS THEORY AND DATA REDUCTION IN INFORMATION SYSTEMS AND DATA MINING ROUGH SETS THEORY AND DATA REDUCTION IN INFORMATION SYSTEMS AND DATA MINING Mofreh Hogo, Miroslav Šnorek CTU in Prague, Departement Of Computer Sciences And Engineering Karlovo Náměstí 13, 121 35 Prague

More information

Influence of knn-based Load Forecasting Errors on Optimal Energy Production

Influence of knn-based Load Forecasting Errors on Optimal Energy Production Influence of knn-based Load Forecasting Errors on Optimal Energy Production Alicia Troncoso Lora 1, José C. Riquelme 1, José Luís Martínez Ramos 2, Jesús M. Riquelme Santos 2, and Antonio Gómez Expósito

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning Lesson 34 Learning From Observations 12.2 Concept Learning Definition: The problem is to learn a function mapping examples into two classes: positive and negative. We are given

More information

Improving the Expert Networks of a Modular Multi-Net System for Pattern Recognition

Improving the Expert Networks of a Modular Multi-Net System for Pattern Recognition Improving the Expert Networks of a Modular Multi-Net System for Pattern Recognition Mercedes Fernández-Redondo 1, Joaquín Torres-Sospedra 1 and Carlos Hernández-Espinosa 1 Departamento de Ingenieria y

More information

Research Article Special Approach to Near Set Theory

Research Article Special Approach to Near Set Theory Mathematical Problems in Engineering Volume 2011, Article ID 168501, 10 pages doi:10.1155/2011/168501 Research Article Special Approach to Near Set Theory M. E. Abd El-Monsef, 1 H. M. Abu-Donia, 2 and

More information

kind include Information Gain (Quinlan, 1994), Gain 1995), and Gini Index (Breiman et al., 1984). dierences in bias.

kind include Information Gain (Quinlan, 1994), Gain 1995), and Gini Index (Breiman et al., 1984). dierences in bias. A Quantication Of Distance-Bias Between Evaluation Metrics In Classication Ricardo Vilalta IBM T.J. Watson Research Center, 30 Saw Mill River Rd., Hawthorne, N.Y., 10532 USA Daniel Oblinger IBM T.J. Watson

More information

Rough Sets, Rough Relations and Rough Functions. Zdzislaw Pawlak. Warsaw University of Technology. ul. Nowowiejska 15/19, Warsaw, Poland.

Rough Sets, Rough Relations and Rough Functions. Zdzislaw Pawlak. Warsaw University of Technology. ul. Nowowiejska 15/19, Warsaw, Poland. Rough Sets, Rough Relations and Rough Functions Zdzislaw Pawlak Institute of Computer Science Warsaw University of Technology ul. Nowowiejska 15/19, 00 665 Warsaw, Poland and Institute of Theoretical and

More information

Banacha Warszawa Poland s:

Banacha Warszawa Poland  s: Chapter 12 Rough Sets and Rough Logic: A KDD Perspective Zdzis law Pawlak 1, Lech Polkowski 2, and Andrzej Skowron 3 1 Institute of Theoretical and Applied Informatics Polish Academy of Sciences Ba ltycka

More information

Index. Cambridge University Press Relational Knowledge Discovery M E Müller. Index. More information

Index. Cambridge University Press Relational Knowledge Discovery M E Müller. Index. More information s/r. See quotient, 93 R, 122 [x] R. See class, equivalence [[P Q]]s, 142 =, 173, 164 A α, 162, 178, 179 =, 163, 193 σ RES, 166, 22, 174 Ɣ, 178, 179, 175, 176, 179 i, 191, 172, 21, 26, 29 χ R. See rough

More information

Decision Tree Learning

Decision Tree Learning Decision Tree Learning Goals for the lecture you should understand the following concepts the decision tree representation the standard top-down approach to learning a tree Occam s razor entropy and information

More information

The Fourth International Conference on Innovative Computing, Information and Control

The Fourth International Conference on Innovative Computing, Information and Control The Fourth International Conference on Innovative Computing, Information and Control December 7-9, 2009, Kaohsiung, Taiwan http://bit.kuas.edu.tw/~icic09 Dear Prof. Yann-Chang Huang, Thank you for your

More information

Enhancing Generalization Capability of SVM Classifiers with Feature Weight Adjustment

Enhancing Generalization Capability of SVM Classifiers with Feature Weight Adjustment Enhancing Generalization Capability of SVM Classifiers ith Feature Weight Adjustment Xizhao Wang and Qiang He College of Mathematics and Computer Science, Hebei University, Baoding 07002, Hebei, China

More information

Dynamic Programming Approach for Construction of Association Rule Systems

Dynamic Programming Approach for Construction of Association Rule Systems Dynamic Programming Approach for Construction of Association Rule Systems Fawaz Alsolami 1, Talha Amin 1, Igor Chikalov 1, Mikhail Moshkov 1, and Beata Zielosko 2 1 Computer, Electrical and Mathematical

More information

Rough Set Model Selection for Practical Decision Making

Rough Set Model Selection for Practical Decision Making Rough Set Model Selection for Practical Decision Making Joseph P. Herbert JingTao Yao Department of Computer Science University of Regina Regina, Saskatchewan, Canada, S4S 0A2 {herbertj, jtyao}@cs.uregina.ca

More information

Some remarks on conflict analysis

Some remarks on conflict analysis European Journal of Operational Research 166 (2005) 649 654 www.elsevier.com/locate/dsw Some remarks on conflict analysis Zdzisław Pawlak Warsaw School of Information Technology, ul. Newelska 6, 01 447

More information

Discriminant Uncorrelated Neighborhood Preserving Projections

Discriminant Uncorrelated Neighborhood Preserving Projections Journal of Information & Computational Science 8: 14 (2011) 3019 3026 Available at http://www.joics.com Discriminant Uncorrelated Neighborhood Preserving Projections Guoqiang WANG a,, Weijuan ZHANG a,

More information

Rough operations on Boolean algebras

Rough operations on Boolean algebras Rough operations on Boolean algebras Guilin Qi and Weiru Liu School of Computer Science, Queen s University Belfast Belfast, BT7 1NN, UK Abstract In this paper, we introduce two pairs of rough operations

More information

A Swarm-based Rough Set Approach for Group Decision Support Systems

A Swarm-based Rough Set Approach for Group Decision Support Systems 2009 Ninth International Conference on Hybrid Intelligent Systems A Swarm-based Rough Set Approach for Group Decision Support Systems Mingyan Zhao 1, Hongbo Liu 2,3, Ajith Abraham 2,3 and Emilio Corchado

More information

Ensembles. Léon Bottou COS 424 4/8/2010

Ensembles. Léon Bottou COS 424 4/8/2010 Ensembles Léon Bottou COS 424 4/8/2010 Readings T. G. Dietterich (2000) Ensemble Methods in Machine Learning. R. E. Schapire (2003): The Boosting Approach to Machine Learning. Sections 1,2,3,4,6. Léon

More information

hsnim: Hyper Scalable Network Inference Machine for Scale-Free Protein-Protein Interaction Networks Inference

hsnim: Hyper Scalable Network Inference Machine for Scale-Free Protein-Protein Interaction Networks Inference CS 229 Project Report (TR# MSB2010) Submitted 12/10/2010 hsnim: Hyper Scalable Network Inference Machine for Scale-Free Protein-Protein Interaction Networks Inference Muhammad Shoaib Sehgal Computer Science

More information

Data Mining and Machine Learning (Machine Learning: Symbolische Ansätze)

Data Mining and Machine Learning (Machine Learning: Symbolische Ansätze) Data Mining and Machine Learning (Machine Learning: Symbolische Ansätze) Learning Individual Rules and Subgroup Discovery Introduction Batch Learning Terminology Coverage Spaces Descriptive vs. Predictive

More information

Rough Sets for Uncertainty Reasoning

Rough Sets for Uncertainty Reasoning Rough Sets for Uncertainty Reasoning S.K.M. Wong 1 and C.J. Butz 2 1 Department of Computer Science, University of Regina, Regina, Canada, S4S 0A2, wong@cs.uregina.ca 2 School of Information Technology

More information

A Simple Implementation of the Stochastic Discrimination for Pattern Recognition

A Simple Implementation of the Stochastic Discrimination for Pattern Recognition A Simple Implementation of the Stochastic Discrimination for Pattern Recognition Dechang Chen 1 and Xiuzhen Cheng 2 1 University of Wisconsin Green Bay, Green Bay, WI 54311, USA chend@uwgb.edu 2 University

More information

Three-Way Analysis of Facial Similarity Judgments

Three-Way Analysis of Facial Similarity Judgments Three-Way Analysis of Facial Similarity Judgments Daryl H. Hepting, Hadeel Hatim Bin Amer, and Yiyu Yao University of Regina, Regina, SK, S4S 0A2, CANADA hepting@cs.uregina.ca, binamerh@cs.uregina.ca,

More information

Decision Tree Learning Lecture 2

Decision Tree Learning Lecture 2 Machine Learning Coms-4771 Decision Tree Learning Lecture 2 January 28, 2008 Two Types of Supervised Learning Problems (recap) Feature (input) space X, label (output) space Y. Unknown distribution D over

More information

Neural Networks and the Back-propagation Algorithm

Neural Networks and the Back-propagation Algorithm Neural Networks and the Back-propagation Algorithm Francisco S. Melo In these notes, we provide a brief overview of the main concepts concerning neural networks and the back-propagation algorithm. We closely

More information

Concept Lattices in Rough Set Theory

Concept Lattices in Rough Set Theory Concept Lattices in Rough Set Theory Y.Y. Yao Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 E-mail: yyao@cs.uregina.ca URL: http://www.cs.uregina/ yyao Abstract

More information

Outlier Detection Using Rough Set Theory

Outlier Detection Using Rough Set Theory Outlier Detection Using Rough Set Theory Feng Jiang 1,2, Yuefei Sui 1, and Cungen Cao 1 1 Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences,

More information

Comparison of Shannon, Renyi and Tsallis Entropy used in Decision Trees

Comparison of Shannon, Renyi and Tsallis Entropy used in Decision Trees Comparison of Shannon, Renyi and Tsallis Entropy used in Decision Trees Tomasz Maszczyk and W lodzis law Duch Department of Informatics, Nicolaus Copernicus University Grudzi adzka 5, 87-100 Toruń, Poland

More information

Drawing Conclusions from Data The Rough Set Way

Drawing Conclusions from Data The Rough Set Way Drawing Conclusions from Data The Rough et Way Zdzisław Pawlak Institute of Theoretical and Applied Informatics, Polish Academy of ciences, ul Bałtycka 5, 44 000 Gliwice, Poland In the rough set theory

More information

CSCI 5622 Machine Learning

CSCI 5622 Machine Learning CSCI 5622 Machine Learning DATE READ DUE Mon, Aug 31 1, 2 & 3 Wed, Sept 2 3 & 5 Wed, Sept 9 TBA Prelim Proposal www.rodneynielsen.com/teaching/csci5622f09/ Instructor: Rodney Nielsen Assistant Professor

More information

ON COMBINING PRINCIPAL COMPONENTS WITH FISHER S LINEAR DISCRIMINANTS FOR SUPERVISED LEARNING

ON COMBINING PRINCIPAL COMPONENTS WITH FISHER S LINEAR DISCRIMINANTS FOR SUPERVISED LEARNING ON COMBINING PRINCIPAL COMPONENTS WITH FISHER S LINEAR DISCRIMINANTS FOR SUPERVISED LEARNING Mykola PECHENIZKIY*, Alexey TSYMBAL**, Seppo PUURONEN* Abstract. The curse of dimensionality is pertinent to

More information

Decision Tree Learning

Decision Tree Learning Decision Tree Learning Goals for the lecture you should understand the following concepts the decision tree representation the standard top-down approach to learning a tree Occam s razor entropy and information

More information

PAC-learning, VC Dimension and Margin-based Bounds

PAC-learning, VC Dimension and Margin-based Bounds More details: General: http://www.learning-with-kernels.org/ Example of more complex bounds: http://www.research.ibm.com/people/t/tzhang/papers/jmlr02_cover.ps.gz PAC-learning, VC Dimension and Margin-based

More information

COMS 4771 Introduction to Machine Learning. Nakul Verma

COMS 4771 Introduction to Machine Learning. Nakul Verma COMS 4771 Introduction to Machine Learning Nakul Verma Announcements HW1 due next lecture Project details are available decide on the group and topic by Thursday Last time Generative vs. Discriminative

More information

Active Sonar Target Classification Using Classifier Ensembles

Active Sonar Target Classification Using Classifier Ensembles International Journal of Engineering Research and Technology. ISSN 0974-3154 Volume 11, Number 12 (2018), pp. 2125-2133 International Research Publication House http://www.irphouse.com Active Sonar Target

More information

Additive Preference Model with Piecewise Linear Components Resulting from Dominance-based Rough Set Approximations

Additive Preference Model with Piecewise Linear Components Resulting from Dominance-based Rough Set Approximations Additive Preference Model with Piecewise Linear Components Resulting from Dominance-based Rough Set Approximations Krzysztof Dembczyński, Wojciech Kotłowski, and Roman Słowiński,2 Institute of Computing

More information

I D I A P. Online Policy Adaptation for Ensemble Classifiers R E S E A R C H R E P O R T. Samy Bengio b. Christos Dimitrakakis a IDIAP RR 03-69

I D I A P. Online Policy Adaptation for Ensemble Classifiers R E S E A R C H R E P O R T. Samy Bengio b. Christos Dimitrakakis a IDIAP RR 03-69 R E S E A R C H R E P O R T Online Policy Adaptation for Ensemble Classifiers Christos Dimitrakakis a IDIAP RR 03-69 Samy Bengio b I D I A P December 2003 D a l l e M o l l e I n s t i t u t e for Perceptual

More information

Hierarchical Structures on Multigranulation Spaces

Hierarchical Structures on Multigranulation Spaces Yang XB, Qian YH, Yang JY. Hierarchical structures on multigranulation spaces. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 27(6): 1169 1183 Nov. 2012. DOI 10.1007/s11390-012-1294-0 Hierarchical Structures

More information

Decision tables and decision spaces

Decision tables and decision spaces Abstract Decision tables and decision spaces Z. Pawlak 1 Abstract. In this paper an Euclidean space, called a decision space is associated with ever decision table. This can be viewed as a generalization

More information

Computational Intelligence, Volume, Number, VAGUENES AND UNCERTAINTY: A ROUGH SET PERSPECTIVE. Zdzislaw Pawlak

Computational Intelligence, Volume, Number, VAGUENES AND UNCERTAINTY: A ROUGH SET PERSPECTIVE. Zdzislaw Pawlak Computational Intelligence, Volume, Number, VAGUENES AND UNCERTAINTY: A ROUGH SET PERSPECTIVE Zdzislaw Pawlak Institute of Computer Science, Warsaw Technical University, ul. Nowowiejska 15/19,00 665 Warsaw,

More information

CS 6375 Machine Learning

CS 6375 Machine Learning CS 6375 Machine Learning Decision Trees Instructor: Yang Liu 1 Supervised Classifier X 1 X 2. X M Ref class label 2 1 Three variables: Attribute 1: Hair = {blond, dark} Attribute 2: Height = {tall, short}

More information

A Logical Formulation of the Granular Data Model

A Logical Formulation of the Granular Data Model 2008 IEEE International Conference on Data Mining Workshops A Logical Formulation of the Granular Data Model Tuan-Fang Fan Department of Computer Science and Information Engineering National Penghu University

More information

A Rough Set Interpretation of User s Web Behavior: A Comparison with Information Theoretic Measure

A Rough Set Interpretation of User s Web Behavior: A Comparison with Information Theoretic Measure A Rough et Interpretation of User s Web Behavior: A Comparison with Information Theoretic easure George V. eghabghab Roane tate Dept of Computer cience Technology Oak Ridge, TN, 37830 gmeghab@hotmail.com

More information

Parameters to find the cause of Global Terrorism using Rough Set Theory

Parameters to find the cause of Global Terrorism using Rough Set Theory Parameters to find the cause of Global Terrorism using Rough Set Theory Sujogya Mishra Research scholar Utkal University Bhubaneswar-751004, India Shakti Prasad Mohanty Department of Mathematics College

More information

Convex envelopes, cardinality constrained optimization and LASSO. An application in supervised learning: support vector machines (SVMs)

Convex envelopes, cardinality constrained optimization and LASSO. An application in supervised learning: support vector machines (SVMs) ORF 523 Lecture 8 Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Any typos should be emailed to a a a@princeton.edu. 1 Outline Convexity-preserving operations Convex envelopes, cardinality

More information