Rough Set Approaches for Discovery of Rules and Attribute Dependencies

Similar documents
A new Approach to Drawing Conclusions from Data A Rough Set Perspective

Interpreting Low and High Order Rules: A Granular Computing Approach

Drawing Conclusions from Data The Rough Set Way

A PRIMER ON ROUGH SETS:

ROUGH SET THEORY FOR INTELLIGENT INDUSTRIAL APPLICATIONS

Selected Algorithms of Machine Learning from Examples

Rough Sets, Rough Relations and Rough Functions. Zdzislaw Pawlak. Warsaw University of Technology. ul. Nowowiejska 15/19, Warsaw, Poland.

Naive Bayesian Rough Sets

Easy Categorization of Attributes in Decision Tables Based on Basic Binary Discernibility Matrix

A Generalized Decision Logic in Interval-set-valued Information Tables

Computational Intelligence, Volume, Number, VAGUENES AND UNCERTAINTY: A ROUGH SET PERSPECTIVE. Zdzislaw Pawlak

Concept Lattices in Rough Set Theory

Learning Rules from Very Large Databases Using Rough Multisets

High Frequency Rough Set Model based on Database Systems

Rough Set Model Selection for Practical Decision Making

Classification of Voice Signals through Mining Unique Episodes in Temporal Information Systems: A Rough Set Approach

A Logical Formulation of the Granular Data Model

Three Discretization Methods for Rule Induction

APPLICATION FOR LOGICAL EXPRESSION PROCESSING

Classification Based on Logical Concept Analysis

Data Analysis - the Rough Sets Perspective

Ensembles of classifiers based on approximate reducts

Banacha Warszawa Poland s:

Knowledge Discovery. Zbigniew W. Ras. Polish Academy of Sciences, Dept. of Comp. Science, Warsaw, Poland

A Scientometrics Study of Rough Sets in Three Decades

Comparison of Rough-set and Interval-set Models for Uncertain Reasoning

2 WANG Jue, CUI Jia et al. Vol.16 no", the discernibility matrix is only a new kind of learning method. Otherwise, we have to provide the specificatio

On Improving the k-means Algorithm to Classify Unclassified Patterns

Rough Set Approach for Generation of Classification Rules of Breast Cancer Data

Rough Set Approach for Generation of Classification Rules for Jaundice

An algorithm for induction of decision rules consistent with the dominance principle

Research on Complete Algorithms for Minimal Attribute Reduction

The size of decision table can be understood in terms of both cardinality of A, denoted by card (A), and the number of equivalence classes of IND (A),

Characterizing Pawlak s Approximation Operators

ARPN Journal of Science and Technology All rights reserved.

ROUGH set methodology has been witnessed great success

Andrzej Skowron, Zbigniew Suraj (Eds.) To the Memory of Professor Zdzisław Pawlak

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

ROUGH SETS THEORY AND DATA REDUCTION IN INFORMATION SYSTEMS AND DATA MINING

Rough Sets for Uncertainty Reasoning

Feature Selection with Fuzzy Decision Reducts

Modeling the Real World for Data Mining: Granular Computing Approach

Some remarks on conflict analysis

Alternative Approach to Mining Association Rules

Fuzzy Rough Sets with GA-Based Attribute Division

1 Introduction Rough sets theory has been developed since Pawlak's seminal work [6] (see also [7]) as a tool enabling to classify objects which are on

Parameters to find the cause of Global Terrorism using Rough Set Theory

REDUCTS AND ROUGH SET ANALYSIS

METRIC BASED ATTRIBUTE REDUCTION IN DYNAMIC DECISION TABLES

Discovery of Concurrent Data Models from Experimental Tables: A Rough Set Approach

Minimal Attribute Space Bias for Attribute Reduction

Rough Sets and Conflict Analysis

Rough operations on Boolean algebras

Granularity, Multi-valued Logic, Bayes Theorem and Rough Sets

On Objective Measures of Rule Surprisingness.

Sets with Partial Memberships A Rough Set View of Fuzzy Sets

Discovery of Functional and Approximate Functional Dependencies in Relational Databases

On Rough Set Modelling for Data Mining

Foundations of Classification

Information Refinement and Revision for Medical Expert System Automated Extraction of Hierarchical Rules from Clinical Data

Tolerance Approximation Spaces. Andrzej Skowron. Institute of Mathematics. Warsaw University. Banacha 2, Warsaw, Poland

Issues in Modeling for Data Mining

The Fourth International Conference on Innovative Computing, Information and Control

Model Complexity of Pseudo-independent Models

CRITERIA REDUCTION OF SET-VALUED ORDERED DECISION SYSTEM BASED ON APPROXIMATION QUALITY

An Analysis of Quantitative Measures Associated with Rules

Mining Approximative Descriptions of Sets Using Rough Sets

Research Article Special Approach to Near Set Theory

Bhubaneswar , India 2 Department of Mathematics, College of Engineering and

Computers and Mathematics with Applications

Data mining using Rough Sets

ON SOME PROPERTIES OF ROUGH APPROXIMATIONS OF SUBRINGS VIA COSETS

Abstract. In this paper we present a query answering system for solving non-standard

Short Note: Naive Bayes Classifiers and Permanence of Ratios

Induction of Decision Trees

Action Rule Extraction From A Decision Table : ARED

An Approach to Classification Based on Fuzzy Association Rules

Notes on Rough Set Approximations and Associated Measures

Rough Soft Sets: A novel Approach

Complexity Theory VU , SS The Polynomial Hierarchy. Reinhard Pichler

Outline. Complexity Theory EXACT TSP. The Class DP. Definition. Problem EXACT TSP. Complexity of EXACT TSP. Proposition VU 181.

USING SAT FOR COMBINATIONAL IMPLEMENTATION CHECKING. Liudmila Cheremisinova, Dmitry Novikov

Title: Logical Agents AIMA: Chapter 7 (Sections 7.4 and 7.5)

IN the areas of machine learning, artificial intelligence, as

Index. C, system, 8 Cech distance, 549

Neighborhoods Systems: Measure, Probability and Belief Functions

High Performance Absorption Algorithms for Terminological Reasoning

COMP219: Artificial Intelligence. Lecture 19: Logic for KR

Granular Computing: Granular Classifiers and Missing Values

FINITE PREDICATES WITH APPLICATIONS IN PATTERN RECOGNITION PROBLEMS

COMP219: Artificial Intelligence. Lecture 19: Logic for KR

Applying Domain Knowledge in Association Rules Mining Process First Experience

Outline. Complexity Theory. Introduction. What is abduction? Motivation. Reference VU , SS Logic-Based Abduction

Fuzzy Modal Like Approximation Operations Based on Residuated Lattices

Hypersequent Calculi for some Intermediate Logics with Bounded Kripke Models

Boolean Algebra CHAPTER 15

Approximate Boolean Reasoning: Foundations and Applications in Data Mining

ENTROPIES OF FUZZY INDISCERNIBILITY RELATION AND ITS OPERATIONS

Mathematical Approach to Vagueness

Nested Epistemic Logic Programs

Transcription:

Rough Set Approaches for Discovery of Rules and Attribute Dependencies Wojciech Ziarko Department of Computer Science University of Regina Regina, SK, S4S 0A2 Canada Abstract The article presents an elementary overview of techniques for data analysis and predictive modeling from data using the rough set approach. The specific knowledge discovery-related data analysis problems discussed in the article are the discovery of functional and partial functional dependencies and the discovery of rules in data. The presentation is focused on the application of the basic rough set model to knowledge discovery and does not include the discussion of the application of the extended probabilistic model of rough set for that purpose. Keywords: rough sets, knowledge discovery, data mining, data dependencies, rules. 1 Introduction The theory of rough sets was introduced by Pawlak [1,7,16,17,19-21,32,43] in early eighties. The intuition behind this idea is derived from the simple fact that in real life, when dealing with sets we often have no means of distinguishing individual set elements. The elements may posses some measurable characteristics but in principle, due to limited resolution of our perception mechanism, we can distinguish only classes of elements rather than individuals. Elements within classes are indistinguishable. For example, assume that we have a large database of hospital patient records and that all what we know about each patient is whether patient s body temperature(bt) is LOW, NORMAL or HIGH, whether patients blood pressure(bp) is LOW, NORMAL or HIGH and whether the heart rate (HR) is NORMAL or ABNORMAL. If that s all what we know, then regardless how many records we would have in our database, we will have maximum eighteen classes of patients corresponding to different combinations of values BP, BT and HR, such as for instance the combination (BT=NORMAL, BP=High, HR=ABNORMAL ). Each combination potentially corresponds to thousands of patient records, each of which is not different, as far as blood pressure and body temperature are concerned, from any other record from this group. We say that the patients have been classified into a certain number of categories, or elementary sets in the rough set terminology. Such a classification has several advantages: 1. Reduction in the complexity of information representation (rather than analyzing huge information files about individual patients we deal with relatively small tables corresponding to observed combinations of generalized features). 2. The ability to discover repetitive patterns in data since each combination potentially corresponds to a large number of patients. 3. The ability to reason about relations among features (attributes) of objects occurring in a large data collection based on results of analysis of relatively small and size-constrained tables, referred to as decision tables, representing classifications of the domain of interest. 0 In: W. Klosgen, J. Zytkow (eds) Handbook of Data Mining and Knowledge Discovery, Oxford University Press, 2002, pp. 328-338 1

OBJ S H E C T 1 12.0 132.2 1715 7.0 75 2 18.2 148.0 3015 32.3 130 3 19.0 175.8 1826 11.2 60................ 9555 17.5 199.1 1917 4.0 143 9556 18.0 111.0 2001 17.1 95 9557 19.6 186.6 4222 29.9 152 9558 15.7 103.2 3832 41.1 161 Table 1: Example raw data representing 9558 objects CLASS S H E C T E 1 0 0 1 0 0 E 2 1 0 2 1 1 E 3 1 1 1 0 0 E 4 0 2 1 1 1 E 5 1 2 1 0 1 E 6 1 0 1 0 0 E 7 1 2 2 1 1 E 8 0 0 2 1 1 Table 2: The decision table derived from Table 1 The theory of rough sets provides a collection of methods which can be applied to the analysis and reduction of such tables. The methods are in particular applicable to data mining or knowledge discovery problems but also have other applications such as trainable control and pattern recognition[30,33-44]. For the purpose of illustration, we will use the decision table given in Table 2, which was derived from data collection partially shown in Table 1 by replacing the original attribute values with some discrete ranges. In this table, the column headings S,H,E,C,T are abbreviations for attribute names SIZE, HEIGHT, ENERGY,CURRENT and TEMPERATURE, respectively. The attributes are generalized properties of objects, identified by sequence numbers in the column OBJ of Table 1. The classes of objects with the same combinations of attribute values are identified in Table 2 in the column CLASS. 2 Discovering Attribute Dependencies An important aspect of KDD is the discovery, analysis, and characterization of dependencies among attributes. The problem of discovery of attribute dependencies was studied independently by many researchers[2,4,12]. In this article, we specifically focus on discovering functional and partial functional dependencies using rough set theory. The probabilistic extension of this theory, called variable precision rough set model, allows for discovery of probabilistic dependencies among attributes [39,40] using essentially the same techniques as presented in the rest of this section. In the rough set approach to discovery of dependencies, the subject of the analysis are decision tables derived from the original data rather than the data itself. As demonstrated in the example shown in the Introduction, the decision table is obtained by performing a suitably selected multidimensional mapping converting the original data into finite-valued secondary attribute-value vectors. The definition of the mapping is domain-dependent and often requires significant domain expertise for proper definition. Typically, the values of the secondary attributes correspond to some generalized qualitative categories, such LOW, MEDIUM, HIGH etc. The key issue is to have sufficient amount of representative data points so that classification of the universe of discourse OBJ (i.e. domain of interest containing objects represented by data points accumulated in the database) is complete. The complete classification covers all feasible combinations of values of the generalized secondary attributes. For example, we will assume that the decision table in Table 2 exhausts all feasible combinations of discretized attributes S,H,E,C,T. In the absence of the complete classification of the 2

domain, any conclusions derived from the decision table should be constrained to the accumulated data only. They can be false in the whole domain from which the data were collected. In particular, the dependencies discovered in the decision table will hold in the data on hand, but may not hold in the whole domain, if the decision table is incomplete. In the search for dependencis, we are interested not only in functional dependencies but also in a spectra of partial functional dependencies. In more detail, in the context of the data-derived decision table, we ask the question as to whether there is any functional or partial functional dependency between groups of attributes P A and Q A, where A is a set of all attributes of the decision table. To precisely define the notions of partial functional dependency, deterministic and non-deterministic decision rules, we need to use some elementary ideas of rough sets theory, as presented below[1]. Let R(P ) be the equivalence relation among objects in the universe OBJ. The pair (OBJ, R(P )) will be called an approximation space[1]. Also, let R (P ) be the collection of equivalence classes (also called elementary sets) of R(P ). That is, elements of R (P ) are groups of objects having the same values of attributes belonging to P. The elementary sets reflect our basic knowledge about the domain, the knowledge in the sense of knowing which categories of objects occur in the domain, and being able to classify each object into one of the categories. With this kind of knowledge it is not possible, in general, to construct discriminating description of any arbitrary subset of the domain. What it means in practice is that some concepts can never be learned precisely with the available information. Instead, rough approximations of these concepts can be learned. These approximations are defined below. They are fundamental notions of rough set theory which are necessary to evaluate the quality of information describing objects of the domain of interest, relative to the target set of objects Y. The quality of the information is expressed in these definitions in terms of the ability to form tight lower and upper approximations of the target set Y. The lower approximation in the approximation space (OBJ, R(P )), or alternatively the interior INT (Y ) of an arbitrary subset Y OBJ, is defined as the union of those equivalence classes of R (P ) which are completely contained by Y, i.e., INT (Y ) = {E R (P ) : E Y } The lower approximation characterizes objects which can be classified into Y without any uncertainty, based on the available information. In other words, the lower approximation is a largest set of objects which has discriminating description and is contained in the target set Y. The upper approximation defines objects which possibly belong to the target set. It is the smallest set of objects having discriminating description and containing the target set. The upper approximation of Y, denoted here as as UP P (Y ), is a union of these elementary classes which have some overlap with Y, i.e. UP P (Y ) = {E R (P ) : E Y }. Because the set of the attributes Q corresponds to the partitioning R (Q) of the universe OBJ, the degree K(P, Q) of the deterministic, or functional dependency in the relationship between attribute collections P and Q can be defined as the total relative size of lower approximations of classes of the partition R (Q) in the approximation space (OBJ, R(P )). That is, if R (Q) = {Y 1, Y 2,..., Y m } then m i=1 K(P, Q) = card(int (Y i)) card(obj) where card is a set cardinality. K(P, Q) assumes values in the range [0,1] with K(P, Q) = 1 for functional dependency and K(P, Q) = 0 when no value of Q attributes can be uniquely determined by values of P attributes. K(P, Q) can be interpreted as a proportion of such objects in the domain for which it suffices to know the values of attributes in P to determine the values of attributes in Q. In the practice, we are most often interested in analyzing dependencies where Q contains only a single attribute. For example, it is easy to verify, based on the Table 2, that for P = {SIZE, HEIGHT, ENERGY, CU RREN T } and Q = {T EM P ERAT U RE} the degree of dependency of K(P, Q) is 1. This means that this dependency is functional whereas the dependency between P = {ENERGY, CURRENT } 3

and Q = {T EMP ERAT URE} is only partially functional with K(P, Q) = 0.5. The methodology for the dependency computation can be summarized in the following algorithm. In the algorithm, it is assumed that there is only one finite-valued target attribute. In the case of several target attributes, any subset of them can be treated as a compound single attribute and processed in the same way. Algorithm DEP: Attribute Dependency Computation 1. Construct decision table from the original data according to pre-defined mapping of data records into secondary attribute vectors. 2. For each value V i of the target attribute identify the set of objects (represented by data records) Y i containing that value. 3. Compute the cardinality of the lower approximation of each set Y i. 4. Output the degree of dependency as the ratio of the total size of all lower approximations of the sets Y i to the total number of objects (data records here). In the rough set approach, any full or partial function reflecting dependency occurring in data is represented by the decision table. The decision table plays the same role as the analytical formula for numeric dependencies (functions). The tabular representation can be simplified, to eliminate any redundant attributes, and analyzed to determine the relative significance of the attributes involved in such a dependency. This step would be equivalent to finding a simpler formula representing a numeric function. The simplification of table dependencies is based on the concept of relative reduct of rough sets theory[1], as presented below. The relative reduct of the attribute collection P, with respect to the dependency K(P, Q), is defined as a subset RED(P, Q) P such that: 1. K(RED(P, Q), Q) = K(P, Q), i.e., relative reduct preserves the degree of inter-attribute dependency, and 2. For any attribute a RED(P, Q), K(RED(P, Q) {a}, Q) K(P, Q), that is the relative reduct is a minimal subset with respect to the property 1. A single relative reduct can be computed in linear time with the following attribute elimination procedure. The outcome of the process depends on the pre-determined priority ordering of attributes. Algorithm RRED: Computation of Relative Reduct 1. Consider all attributes of the initial collection of attributes P, one by one, in the reverse order of priority. For each attribute a do the following: 1.1 Remove the attribute a from P temporarily. 1.2 Check if the dependency K(P, Q) changed. If yes, return a back to P, else remove the attribute a from P permanently. 2. Output the set of attributes remaining in P at the end of the process. It is a relative reduct of the initial set of attributes P. For example, one possible relative reduct with respect to the dependency between P = {SIZE, HEIGHT, ENERGY, CURRENT } and Q = {T EMP ERAT URE} is RED(P, Q) = {HEIGHT, EN ERGY }. This means that the discovered dependency can be characterized by fewer attributes, leading to possible savings in information representation, better understanding of the nature of the dependency and stronger patterns. In general, a number of alternative reducts can be computed for each analyzed dependency and the one of the lowest total cost can be selected to represent the discovered dependency. The dependency represented in the reduced form is usually more regular as it reflects stronger data patterns. This is illustrated in Table 3. For computing all attribute reducts, the decision matrix methodology, described in the next section, can be used. The details are provided in [29]. More research results and algorithms related to finding reducts can be also found in [14]. 4

HEIGHT ENERGY TEMPERATURE 0 1 0 0 2 1 1 1 0 2 1 1 2 2 1 Table 3: Reduced representation of the dependency between P and Q To find fundamental factors contributing to the discovered dependency the idea of relative core can be used. The relative core set of attributes with respect to the dependency K(P, Q) is a subset CORE P such that for all a CORE, K(P {a}, Q) K(P, Q). In other words, CORE is a set of the essential attributes which cannot be eliminated from P without affecting the dependency between P and Q. For example, the core of the dependency between P = {SIZE, HEIGHT, ENERGY, CURRENT } and Q = {T EM P ERAT U RE} is {HEIGHT }. That is, HEIGHT is the fundamental dependency factor and it is included in every relative reduct representing this dependency[1]. The relative core is a context sensitive notion (depends on other attributes in the decision table) and can be empty. This can be interpreted as a case of highly over specified system with superfluous attributes. In a reduct, all attributes are core attributes (relative to the reduct attributes only) since the elimination of any attribute from the reduct results in a drop of the dependency. The percentage drop of the dependency can be used as a measure of relative significance of an attribute in the reduct: the higher the drop the more significant is the attribute with respect to the dependency K(RED(P, Q), Q). For instance, in the dependency between reduct attributes RED(P, Q) = {HEIGHT, ENERGY } and the attribute {T EMP ERAT URE}, as represented in the Table 3, the relative significance of HEIGHT is 60 percent whereas the significance of ENERGY is 40 percent. 3 Rule Discovery Discovering rules from data is one of the most important tasks of knowledge discovery in databases. Many systems and approaches for rule computation have been used for rule or decision tree discovery[3-6,9-11,13,24,26,28-29,30-32,40-44]. Rules can be perceived as data patterns which represent relationships between attribute values. In the rough sets approach, we distinguish certain, or deterministic rules, and possible, or non-deterministic rules. In addition to the above, in the variable precision rough set model, the probabilistic rules can be computed[8,15,18,27]. Since the rule computation methodology developed within the original rough set framework is directly applicable to computation of probabilistic rules by using probabilistic definitions of set approximations, in this section we will focus on rule discovery with the original rough set model only. In data mining context, the most useful are the maximally general rules, the ones which minimize the number of rule conditions. We will refer to such rules as minimal rules. They correspond to value reducts known in the rough set theory, or to prime implicants of the digital circuit design theory. The rough set-based algorithms for rule discovery can either compute rules forming minimal cover or approximation of minimal cover of the target class, or they can be used to find all minimal rules for the target class. Representative algorithms of both kinds are presented in the rest of this section. To precisely define the concept of rules in the rough set approach some additional notational conventions are needed. Let V denote a selected value of the decision (target) attribute d A and let supp(d = V ) denote the set of objects matching this value, referred to as target set of objects. The rules can be computed either with respect to lower approximation INT (supp(d = V )) or upper approximation UP P (supp(d = V )) of the target set. When the lower approximation is selected then deterministic rules are obtained. The non-deterministic rules are obtained from upper approximation of the target set. In either case, the computational procedure is the same. Therefore, to describe the methods we can assume, without loss of generality, that supp(d = V ) is an exact set, or not rough, i. e., supp(d = V ) = INT (supp(d = V )) = UP P (supp(d = V )), as otherwise supp(d = V ) would be substituted by either INT (supp(d = V )) or UP P (supp(d = V )). In our notation, the rule derived from data is a logical implication formula corresponding to the combination of values of some attributes such that the set of all objects matching this combination 5

is contained in the target set of objects. The rule r can be expressed as r : (a i1 = V i1 ) (a i2 = V i2 ),... (a in = V in ) (d = V ) where a ij, V ij denote condition attributes and attribute values, respectively. The set of attributevalue pairs occurring on the left hand side of the rule r is referred to as the rule condition part, denoted cond(r), and the right hand side is the decision part, dec(r), so the rule can be expressed as cond(r) dec(r). The set of all objects supp(cond(r)) in the universe OBJ whose attribute values match the rule conditions is called the rule support. To define minimal rules, the notion of value reduct is introduced first. The value reduct of the set of attribute-value pairs cond(r) is a subset red(r) cond(r) such that 1. supp(red(r)) supp(d = V ) i.e., value reduct preserves the relation of inclusion of the rule support set in the target set, and 2. For any attribute-value pair (a ij = V ij ) red(r), supp(red(r) {(a ij = V ij )}) supp(d = V ) that is, the value reduct is a minimal subset with respect to the property 1. Similar to relative reduct, a single value reduct can be computed in linear time with the following value elimination procedure. The outcome of the process depends on the pre-determined priority ordering of attributes. Algorithm VRED: Computation of Value Reduct 1. Consider all attribute-value pairs of the set cond(r), one by one, in the reverse order of priority. For each pair x do the following: 1.1 Remove the pair x from cond(r) temporarily. 1.2 Check if the support set of cond(r), supp(cond(r)) is included in supp(d = V ). If no, return x back to cond(r), else remove the pair x from cond(r) permanently. 2. Output the set of attribute-value pairs remaining in cond(r) at the end of the process. It is a value reduct of the original contents of cond(r). Every value reduct corresponds to a minimal rule red(r) dec(r) derived from the initial rule cond(r) dec(r). For example, some of the value reducts of the set of attribute-value pairs of the rule corresponding to the first row of the Table 2, with respect to target value T = 0, are {(H = 0), (E = 1)} and {(S = 0), (C = 0)}. These value reducts translate into rules such as, for example the following: (HEIGHT = 0) (ENERGY = 1) (T EMP ERAT URE = 0 ) (SIZE = 0) (CURRENT = 0) (T EMP ERAT URE = 0 ) As it follows from the above, a single minimal rule can be computed in linear time in the number of attributes. Some minimal rules, selected according to a predefined heuristic and covering the target set, or rough approximation of it, can be computed relatively easily in time proportional to number of attributes and number of elementary sets in lower(or upper) approximation of the target class. Algorithms for computing such rules have been implemented in several systems for machine learning and data mining [15,25,29,30,34,41-44]. A representative simplified algorithm is presented below. It is aimed at generating close to minimum covering of the lower (or upper) approximation of the target class with the support sets of selected value reducts. Algorithm MinRul: Computation of Minimal Rules 1. Compute lower (or upper) approximation of the target class. Initialize the list L of value reducts to empty. 6

i j CLASS S H E C T - 1 E 1 0 0 1 0 0 1 - E 2 1 0 2 1 1-2 E 3 1 1 1 0 0 2 - E 4 0 2 1 1 1 3 - E 5 1 2 1 0 1-3 E 6 1 0 1 0 0 4 - E 7 1 2 2 1 1 5 - E 8 0 0 2 1 1 Table 4: The decision table with additional class indexes 2. In the decision table, consider one by one every row corresponding to the selected lower (or upper) approximation of the target set. Compute a single value reduct from the current row. If the support set of the computed value reduct is contained in the union of support sets of the value reducts in the list L, go to the next row; otherwise add the computed value reduct to the list L. 3. Output the list L. At the end of the process, the list L will contain non-redundant set of value reducts whose support sets will cover the selected approximation of the target set. These reducts will directly correspond to the set of rules discriminating either the lower, or the upper approximation of the target class. For illustration, let s compute minimal rules for T = 1 from the decision table shown in Table 2. The lower approximation of supp(t = 1) is the union of elementary classes E 2, E 4, E 5, E 7, E 8. Consequently, we will compute value reducts for rows 2,4,5,7,8 of the decision table using VRED algorithm (the assumed order of value elimination is from left to right). It can be easily verified that the value reduct of rows 2,4,7 and 8 is {(C = 1)}. The value reduct of row 5 is {(H = 2)}. Since the support set of {(H = 2)} is not covered by the support set of {(C = 1)} the result consist of the value reducts {(C = 1)} and {(H = 2)}, and corresponding to them rules: (CURRENT = 1) (T EMP ERAT URE = 1 ) (HEIGHT = 2) (T EMP ERAT URE = 1 ) For knowledge discovery applications, finding all possible minimal rules rather than minimal covering rules is more desirable. However, the problem of computing all minimal rules is NP-hard. All minimal rules can be computed, if the problem size permits, using the techniques of discernibility matrix [10], or decision matrix [9] which is described below. Before we define the concept of a decision matrix, we will assume some notational conventions. That is, we will assume that all classes containing objects belonging to supp(d = V ) and all classes with objects belonging to the complement of supp(d = V ) are separately numbered with subscripts i (i = 1, 2,...γ) and j (j = 1, 2,...ρ) respectively. The modified example decision table with additional indexes is shown in Table 4. A decision matrix M(S) = (M ij ) of a decision table with respect to value V of the decision attribute d is defined as a matrix whose entries are sets of attribute value pairs M ij = {(a, a(e i )) : a(e i ) a(e j )} where a(e i ) denotes the common value of the attribute a on all objects belonging to the class E i. The set M ij contains all attribute-value pairs (attribute, value) whose values are not identical between E i and E j. In other words, M ij represents the complete information distinguishing E i from E j. For example, the entry M 11 of the decision matrix given in Table 5 reflects the fact that class E 1 differs from the class E 2 in values of attributes SIZE, ENERGY, and CURRENT. These values are SIZE = 1, ENERGY = 2 and CURRENT = 1 for the class E 2. The set of all minimal value reducts of the collection of attribute value pairs corresponding to row i of the decision table, and consequently the set of all minimal decision rules for that row can be obtained by forming the Boolean expression 7

j 1 2 3 i CLASS E 1 E 3 E 6 1 E 2 (S,1)(E,2) (H,0)(E,2) (E,2)(C,1) (C,1) (C,1) 2 E 4 (H,2)(C,1) (S,0)(H,2) (S,0)(H,2) (C,1) (C,1) 3 E 5 (S,1)(H,2) (H,2) (H,2) 4 E 7 (S,1)(H,2) (H,2)(E,2) (H,2)(E,2) (E,2)(C,1) (C,1) (C,1) 5 E 8 (E,2)(C,1) (S,0)(H,0) (S,0)(E,2) (E,2)(C,1) (C,1) Table 5: Decision matrix for T EMP ERAT URE = 1 of Table 3. B i = j Mij where and are respectively generalized conjunction and disjunction operators[9,10]. The Boolean expression, called a decision function B i is constructed out of row i of the decision matrix, that is (M i1,m i2,...m iρ ), by formally treating each attribute-value pair occurring in the matrix entry M ij as a Boolean variable and then forming a Boolean conjunction of disjunctions of attribute-value pairs belonging to each set M ij. The decision rules are obtained by turning each decision function into disjunctive normal form and using the absorption law of Boolean algebra to simplify it. The conjuncts, or prime implicants of the simplified decision function correspond to the minimal decision rules [9,10]. The major steps in computing all minimal rules for the selected approximation (lower or upper) of the target class are summarized in the following algorithm. Algorithm AllRul: Computation of All Minimal Rules 1. Split the rows of the input decision table into two categories: lower (or upper) approximation rows of the target set versus all other rows. 2. Construct the decision matrix by comparing decision table rows belonging to the categories obtained in Step 1. 3. Form decision functions out of rows of the decision matrix and turn them into disjunctive normal form. 4. Identify prime implicants of the simplified decision functions obtained in Step 3. 5. Perform the set union operation on the sets of prime implicants associated with the rows of the decision matrix. 6. Output minimal rules for lower (or upper) approximation of the target class by translating all prime implicants obtained in Step 5 into rule format. Each prime implicant is translated into rule conditions with the rule conclusion corresponding to the selected value V of the decision attribute. For example, for the target class corresponding to T EMP ERAT URE = 1 of the decision table shown in Table 4, we can compute and simplify the following decision functions, as derived from the decision matrix presented in Table 5. B 1 = ((S, 1) (E, 2) (C, 1)) ((H, 0) (E, 2) (C, 1)) ((E, 2) (C, 1)) = (E, 2) (C, 1) B 2 = ((H, 2) (C, 1)) ((S, 0) (H, 2) (C, 1)) ((S, 0) (H, 2) (C, 1)) = (H, 2) (C, 1) B 3 = ((S, 1) (H, 2)) ((H, 2)) ((H, 2)) = (H, 2) B 4 = ((S, 1) (H, 2) (E, 2) (C, 1)) ((H, 2) (E, 2) (C, 1)) 8

((H, 2) (E, 2) (C, 1)) = (H, 2) (E, 2) (C, 1) B 5 = ((E, 2) (C, 1)) ((S, 0) (H, 0) (E, 2) (C, 1)) ((S, 0) (E, 2) (C, 1)) = (E, 2) (C, 1) The prime implicants are (E, 2), (C, 1) and (H, 2). After translation, the following rules are obtained: (ENERGY = 2) (T EMP ERAT URE = 1 ) (CURRENT = 1) (T EMP ERAT URE = 1 ) (HEIGHT = 2) (T EMP ERAT URE = 1 ) All the minimal rules for the decision class 0 can be computed in a similar way. The level of difficulty with applying the decision matrix method in practical situations depends on the complexity of the decision table obtained through classification of the original data based on values of generalized secondary attributes. Typically, the attributes used to construct a decision table are the functions of the original data values, their number is small and each attribute has a small number of domain values. This leads to relatively small and constrained number of classes of the decision table. The size of the decision table is constrained by the product of cardinalities of attribute domains and is independent of the size of the original data set. Consequently, careful definition of the decision table attributes is an essential prerequisite before applying the decision matrix method and, in fact, any other rough set-based rule discovery method. On the other hand, the computation of prime implicants associated with decision functions, the most time consuming aspect of the algorithm AllRul, can be done independently, for each row of the decision matrix. This means that the decision matrix method can take full advantage of a parallel implementation in a multiprocessor system to speed up the rule computation process. 4 Final Remarks The theory of rough sets is a fundamental mathematical model which has been extensively studied by mathematicians, logicians and computer scientists. The simplicity and mathematical clarity of the model makes it attractive for both theoreticians and application-oriented researchers. Although its introduction was not motivated by any particular application, it touched the essence of many old problems in AI, pattern recognition and control. More recently, it became clear that rough set theory provides a framework for studying data mining and for developing data analysis algorithms for this research area. However, it should be emphasized that, while providing sound theoretical basis for designing new algorithms, the rough set theory is not associated with any particular method or algorithm for data mining or other applications. The choice of rough set-based techniques for data mining presented in this article reflects only author s preferences and experiences and by no means pretends to cover the whole subject area. 5 Acknowledgments The research reported in this article was partially supported by a research grant awarded to the author by Natural Sciences and Engineering Research Council of Canada. Many thanks to Jan Zytkow and the anonymous referees for precise and inspiring comments. References [1] Pawlak, Z. (1991) Rough Sets: Theoretical Aspects of Reasoning About Data, Kluwer Academic Publishers. [2] Piatetsky-Shapiro, G. (ed.) (1993) Proc. of AAAI-93 Workshop on Knowledge Discovery in Databases, Washington D.C.. [3] Ziarko, W. (ed.) (1994) Rough Sets, Fuzzy Sets and Knowledge Discovery, Springer-Verlag. 9

[4] Piatetsky-Shapiro, G. and Frawley, W. J. (eds.) (1991) Knowledge Discovery in Databases, AAAI/MIT Press. [5] Hu, X., Cercone, N., J. Han (1994) An Attribute-Oriented Rough Set Approach for Knowledge Discovery in Databases, in Ziarko, W. (ed.) Rough Sets, Fuzzy Sets and Knowledge Discovery, Springer-Verlag, pp.90-99. [6] Piatetsky-Shapiro, G. (1994) Knowledge Discovery in Databases: Progress and Challenges,in Ziarko, W. (ed.) Rough Sets, Fuzzy Sets and Knowledge Discovery, Springer-Verlag, pp. 1-10. [7] Pawlak, Z. and Orlowska, E. (1984) Expressive Power of Knowledge Representation, International Journal of Man-Machine Studies, 20, pp. 485-500. [8] Katzberg, J. and Ziarko, W. (1994) Variable Precision Rough Sets with Asymmetric Bounds, in Ziarko, W. (ed.) Rough Sets, Fuzzy Sets and Knowledge Discovery, Springer-Verlag, pp. 167-177. [9] Ziarko, W. and Shan, N. (1996) A Method for Computing All Maximally General Rules in Attribute-Value Systems, Computational Intelligence: An International Journal, vol. 12, no. 2, pp. 223-234. [10] Skowron, A. and Rauszer, C. (1992) The Discernibility Matrices and Functions in Information Systems, in Slowinski, R. (ed.) Intelligent Decision Support: Handbook of Applications and Advances of Rough Sets Theory, Kluwer. [11] Piatetsky-Shapiro, G. (1991) Discovery, Analysis, and Presentation of Strong Rules, in Piatetsky-Shapiro, G. Frawley, W. (eds). Knowledge Discovery in Databases, AAAI/MIT Press, pp. 229-248. [12] Schlimmer, J.V. (1993) Using Learned Dependencies to Automatically Construct Sufficient and Feasible Editing Views, Proceeding of AAAI Workshop on Knowledge Discovery in Databases, Washington D.C. pp. 186-196. [13] Major, J. and Mangano, J. (1993) Selecting Among Rules Induced from a Hurricane Database, Proceeding of AAAI Workshop on Knowledge Discovery in Databases, Washington D.C., pp.28-41. [14] Kryszkiewicz, M., Rybinski, H. (1994) Finding Reducts in Composed Information Systems, In Ziarko, W. (ed.) Rough Sets, Fuzzy Sets and Knowledge Discovery, Springer-Verlag, pp. 261-274. [15] Ziarko, W. and Shan, N. (1994) KDD-R: A Comprehensive System for Knowledge Discovery in Databases Using Rough Sets, Proceedings of the International Workshop on Rough Sets and Soft Computing, RSSC 94, San Jose. [16] Pawlak, Z. (1982) Rough Sets, International Journal of Information and Computer Sciences, vol. 11, no. 5, pp. 341-356. [17] Pawlak, Z. (1984) Rough Classification, International Journal of Man-Machine Studies, 20, pp. 469-483. [18] Ziarko, W. (1993) Variable Precision Rough Sets Model, Journal of Computer and Systems Sciences, vol. 46. no. 1, pp. 39-59. [19] Pawlak, Z., Wong, S.K.M., Ziarko, W. (1988) Rough Sets: Probabilistic versus Deterministic Approach, International Journal of Man-Machine Studies, vol. 29, pp. 81-95. [20] Marek, W., Pawlak, Z. (1984) Rough Sets and Information Systems, Fundamenta Informaticae, 17, 1, pp. 105-115. [21] Wasilewska, A. Syntactic Decision Procedures in Information Systems, International Journal of Man-Machine Studies, 30, pp. 273-285. [22] Wong, S.K.M. and Ziarko, W. (1986) Comparison of the Probabilistic Approximate Classification and the Fuzzy Set Model, International Journal for Fuzzy Sets and Systems, 21, pp. 357-362. 10

[23] 23) Pawlak, Z. (1994) Rough Sets Present State and Further Prospects, in Lin, T.Y. and Wildberger A.M. (eds.), Proceedings of the Third International Workshop on Rough Sets and Soft Computing, San Jose, pp. 3-5. [24] Gawrys, M. and Sienkiewicz, J. (1994) RSL-The Rough Set Library version 2.0, ICS Research Report 27/94, Institute of Computer Science, W.U. of T, Warsaw, Poland. [25] Grzymala-Busse, J. W. (1992). LERS - a System Learning from Examples Based on Rough Sets, in Slowinski, R. (ed.) Intelligent Decision Support: Handbook of Applications and Advances of Rough Sets Theory, Kluwer. [26] Grzymala-Busse, D. and Grzymala-Busse, J. (1994). Comparison of Machine Learning and Knowledge Acquisition Methods of Rule Induction Based on Rough Sets,In Ziarko, W. (ed.) Rough Sets, Fuzzy Sets and Knowledge Discovery, Springer-Verlag, pp. 282-289. [27] Katzberg, J. and Ziarko, W. (1994) Variable Precision Rough Sets with Asymmetric Bounds, in Ziarko, W. (ed.) Rough Sets, Fuzzy Sets and Knowledge Discovery, Springer-Verlag, pp. 167-177. [28] Lenarcik, A. and Piasta, Z. (1993). Probabilistic Approach to Decision Algorithm Generation in the Case of Continuous Condition Attributes, Foundations of Computing and Decision Sciences, vol. 18, no. 3-4, pp. 213-224. [29] Reduct Systems Inc. (1994) DATAQUEST and DATALOGIC manual of the software product, Canada. [30] Shan, N. and Ziarko, W. (1995) Data-Based Acquisition and Incremental Modification of Classification Rules, Computational Intelligence: An International Journal, Special Issue on Rough Sets and Knowledge Discovery, vol. 11, no. 2, pp. 357-370. [31] Slowinski, R. and Stefanowski, J. (1992) ROUGHDAS and ROUGHCLASS software Implementation of the Rough Sets Approach, in Slowinski, R. (ed.) Intelligent Decision Support: Handbook of Applications and Advances of Rough Sets Theory, Kluwer. [32] Fayyad, U. and Uthurusamy, R. (eds.) (1995) Proceedings of the First International Conference on Knowledge Discovery and Data Mining, Montreal, Canada 1995. [33] Pawlak, Z. Grzymala-Busse, J. Slowinski, R. and Ziarko, W. (1995) Rough Sets, Communications of The ACM, vol. 38, no. 11, pp. 89-95. [34] Plonka, L. and Mrozek, A. (1995) Rule-Based Stabilization of The Inverted Pendulum, Computational Intelligence: An International Journal, Special Issue on Rough Sets and Knowledge Discovery, vol. 11, no. 2, pp. 348-356 [35] Lin, T.Y.(ed.) 1996 Special Issue on Rough Sets and Soft Computing, Intelligent Automation and Soft Computing: An International Journal, vol. 2, no. 2. [36] Ziarko, W. (ed.) (1996) Special Issue on Rough Sets, Fundamenta Informaticae, vol. 27, no. 2-3. [37] Tsumoto, S. Shan, N. Ziarko, W. and Tanaka, H. (1995) Knowledge Discovery in Clinical Databases Based on Variable Precision Rough Sets, Journal of American Medical Informatics Association, pp. 270-274. [38] Munakata, T. (1998) Fundamentals of the New Artificial Intelligence: Beyond Traditional Paradigms, Springer Verlag. [39] Ziarko, W. (1998) Approximation Region-based Decision Tables, in: Polkowski, L., Skowron, A. (eds.) Rough Sets and Current Trends in Computing, Lecture Notes in AI 1424, Springer Verlag, 178-185. [40] Ziarko, W. (1999) Predictive Modelling with Probabilistic Decision Tables, Proceedings of the 18th IASTED Conference on Modelling, Identification and Control, Innsbruck, 59-62. [41] Polkowski, L., Skowron, A. (eds.) (1998) Rough Sets in Knowledge Discovery, Physica Verlag, vol. 1-2. 11

[42] Son, N. (1997) Rule Induction from Continuous Data, in: Wang, P.(ed.), Proceedings of the Joint Conference of Information Sciences. Duke University, vol. 3, 81-84. [43] Zhong, N., Dang, J., Ohsuga, S. (1998). Soft techniques to data mining. In: Polkowski, L., Skowron, A. (eds.). Rough Sets and Current Trends in Computing. Lecture Notes in AI 1424, Springer Verlag, 231-238. [44] Munakata, T. (1998). Fundamentals of the New Artificial Intelligence. Springer Verlag. [45] Tsumoto, S. (1998). Formalization and induction of medical expert system rules based on rough set theory. In: Polkowski, L., Skowron, A. (eds.) (1998). Rough Sets in Knowledge Discovery. Physica Verlag, vol. 2,.307-323. 12