Similarity-based Classification with Dominance-based Decision Rules

Similar documents
An algorithm for induction of decision rules consistent with the dominance principle

Dominance-based Rough Set Approach Data Analysis Framework. User s guide

Additive Preference Model with Piecewise Linear Components Resulting from Dominance-based Rough Set Approximations

Relationship between Loss Functions and Confirmation Measures

Statistical Model for Rough Set Approach to Multicriteria Classification

Roman Słowiński. Rough or/and Fuzzy Handling of Uncertainty?

Interpreting Low and High Order Rules: A Granular Computing Approach

Feature Selection with Fuzzy Decision Reducts

The possible and the necessary for multiple criteria group decision

Stochastic Dominance-based Rough Set Model for Ordinal Classification

Non-additive robust ordinal regression with Choquet integral, bipolar and level dependent Choquet integrals

Naive Bayesian Rough Sets

ROBUST ORDINAL REGRESSION

A new Approach to Drawing Conclusions from Data A Rough Set Perspective

Additive Consistency of Fuzzy Preference Relations: Characterization and Construction. Extended Abstract

Intuitionistic Fuzzy Sets - An Alternative Look

Computational Intelligence, Volume, Number, VAGUENES AND UNCERTAINTY: A ROUGH SET PERSPECTIVE. Zdzislaw Pawlak

Semantic Rendering of Data Tables: Multivalued Information Systems Revisited

Classification of Voice Signals through Mining Unique Episodes in Temporal Information Systems: A Rough Set Approach

Some remarks on conflict analysis

Fuzzy Systems. Introduction

Rough Set Model Selection for Practical Decision Making

Reference-based Preferences Aggregation Procedures in Multicriteria Decision Making

An Evaluation of the Reliability of Complex Systems Using Shadowed Sets and Fuzzy Lifetime Data

Fuzzy Systems. Introduction

On (Weighted) k-order Fuzzy Connectives

Andrzej Skowron, Zbigniew Suraj (Eds.) To the Memory of Professor Zdzisław Pawlak

AGGREGATION OPERATORS FOR MULTICRITERIA DECISION AID. Jean-Luc Marichal University of Liège

A characterization of the 2-additive Choquet integral

Interactive Multi-Objective Optimization (MOO) using a Set of Additive Value Functions

Rough Sets, Rough Relations and Rough Functions. Zdzislaw Pawlak. Warsaw University of Technology. ul. Nowowiejska 15/19, Warsaw, Poland.

Mining Approximative Descriptions of Sets Using Rough Sets

Compenzational Vagueness

An Analysis on Consensus Measures in Group Decision Making

ROUGH set methodology has been witnessed great success

Favoring Consensus and Penalizing Disagreement in Group Decision Making

Notes on Rough Set Approximations and Associated Measures

FUZZY ASSOCIATION RULES: A TWO-SIDED APPROACH

A PRIMER ON ROUGH SETS:

A Generalized Decision Logic in Interval-set-valued Information Tables

Index. C, system, 8 Cech distance, 549

Drawing Conclusions from Data The Rough Set Way

S-MEASURES, T -MEASURES AND DISTINGUISHED CLASSES OF FUZZY MEASURES

A Logical Formulation of the Granular Data Model

Chebyshev Type Inequalities for Sugeno Integrals with Respect to Intuitionistic Fuzzy Measures

Chapter 2 An Overview of Multiple Criteria Decision Aid

Easy Categorization of Attributes in Decision Tables Based on Basic Binary Discernibility Matrix

Granular Computing: Granular Classifiers and Missing Values

A Rough Set Interpretation of User s Web Behavior: A Comparison with Information Theoretic Measure

Foundations of Classification

Preservation of graded properties of fuzzy relations by aggregation functions

Classification Based on Logical Concept Analysis

Modern Information Retrieval

Strategic Manipulation and Regular Decomposition of Fuzzy Preference Relations

Data mining using Rough Sets

Ordinal Classification with Decision Rules

Weighting Expert Opinions in Group Decision Making for the Influential Effects between Variables in a Bayesian Network Model

On measuring and testing the ordinal correlation between valued outranking relations

Ensembles of classifiers based on approximate reducts

Applied Logic. Lecture 3 part 1 - Fuzzy logic. Marcin Szczuka. Institute of Informatics, The University of Warsaw

On the Relation of Probability, Fuzziness, Rough and Evidence Theory

Chi-square goodness-of-fit test for vague data

On rule acquisition in incomplete multi-scale decision tables

Three Discretization Methods for Rule Induction

Rough Sets for Uncertainty Reasoning

Brock University. Probabilistic granule analysis. Department of Computer Science. Ivo Düntsch & Günther Gediga Technical Report # CS May 2008

ON SOME PROPERTIES OF ROUGH APPROXIMATIONS OF SUBRINGS VIA COSETS

Directional Monotonicity of Fuzzy Implications

ELECTRE TRI-C: A Multiple Criteria Sorting Method Based on Central Reference Actions

Left-continuous t-norms in Fuzzy Logic: an Overview

ENSEMBLES OF DECISION RULES

Classification of Ordinal Data Using Neural Networks

Bulletin of the Transilvania University of Braşov Vol 10(59), No Series III: Mathematics, Informatics, Physics, 67-82

Interleaved Alldifferent Constraints: CSP vs. SAT Approaches

ENTROPIES OF FUZZY INDISCERNIBILITY RELATION AND ITS OPERATIONS

Intuitionistic Fuzzy Sets: Spherical Representation and Distances

The local equivalence of two distances between clusterings: the Misclassification Error metric and the χ 2 distance

Multi-Criteria Optimization - an Important Foundation of Fuzzy System Design

Fuzzy Multiple Criteria Decision Making Methods

Application of Rough Set Theory in Performance Analysis

Banacha Warszawa Poland s:

1 Introduction Rough sets theory has been developed since Pawlak's seminal work [6] (see also [7]) as a tool enabling to classify objects which are on

On the VC-Dimension of the Choquet Integral

Minimal Attribute Space Bias for Attribute Reduction

Solving Classification Problems By Knowledge Sets

Credibilistic Bi-Matrix Game

Rough Sets and Conflict Analysis

REDUCTS AND ROUGH SET ANALYSIS

GENERAL AGGREGATION OPERATORS ACTING ON FUZZY NUMBERS INDUCED BY ORDINARY AGGREGATION OPERATORS

Lecture Notes 1: Vector spaces

Advances in the Use of MCDA Methods in Decision-Making

Fuzzy Modal Like Approximation Operations Based on Residuated Lattices

Model Complexity of Pseudo-independent Models

Text Mining. Dr. Yanjun Li. Associate Professor. Department of Computer and Information Sciences Fordham University

Pregroups and their logics

A Combined Approach for Outliers Detection in Fuzzy Functional Dependency through the Typology of Fuzzy Rules

FUZZY DIAGNOSTIC REASONING THAT TAKES INTO ACCOUNT THE UNCERTAINTY OF THE RELATION BETWEEN FAULTS AND SYMPTOMS

Valued relations aggregation with the Borda method.

Scientific/Technical Approach

A Comparative Study of Different Order Relations of Intervals

Transcription:

Similarity-based Classification with Dominance-based Decision Rules Marcin Szeląg, Salvatore Greco 2,3, Roman Słowiński,4 Institute of Computing Science, Poznań University of Technology, 60-965 Poznań, Poland, {mszelag, rslowinski}@cs.put.poznan.pl 2 Department of Economics and Business, University of Catania, Corso Italia, 55, 9529 Catania, Italy, salgreco@unict.it 3 Portsmouth Business School, University of Portsmouth, UK 4 Systems Research Institute, Polish Academy of Sciences, 0-447 Warsaw Abstract. We consider a similarity-based classification problem where a new case (object) is classified based on its similarity to some previously classified cases. In this process of case-based reasoning (CBR), we adopt the Dominance-based Rough Set Approach (DRSA), that is able to handle monotonic relationship the more similar is object y to object x with respect to the considered features, the closer is y to x in terms of the membership to a given decision class X. At the level of marginal similarity concerning single features, we consider this similarity in ordinal terms only. The marginal similarities are aggregated within induced decision rules describing monotonic relationship between comprehensive similarity of objects and their similarities with respect to single features. Key words: Classification, Similarity, Case-based reasoning, Dominancebased Rough Set Approach, Decision rules Introduction People tend to solve new problems using the solutions of similar problems encountered in the past. This process if often referred to as case-based reasoning (CBR) [9]. As observed by Gilboa and Schmeidler [3], the basic idea of CBR can be found in the following sentence of Hume []: From causes which appear similar we expect similar effects. This is the sum of all our experimental conclusions. We can rephrase this sentence by saying: The more similar are the causes, the more similar one expects the effects. We consider classification performed according to the (broadly construed) CBR paradigm, i.e., a similarity-based classification. In the similarity-based classification problem, there is given a finite set of training objects (case base), described by a set of features, a set of marginal similarity functions (one for each feature), and a set of predefined decision classes. This information is used to suggest membership of a new (unseen) object to particular decision classes. In case-based reasoning, one needs a similarity model aggregating marginal similarities into comprehensive similarity. Traditionally, this model has the form

of a real-valued aggregation function (e.g., Euclidean norm) or binary relation (e.g., fuzzy relation). In this paper, we present a method based on the Dominance-based Rough Set Approach (DRSA) [4,, 2], using a new similarity model in terms of a set of if-then decision rules employing dominance relation in the space created by marginal similarity functions. The first method concerning application of DRSA to CBR was introduced in [5 7], and then extended in [4]. The method presented in this paper, first described in an unpublished PhD thesis [3], concerns revision and improvement of the approach given in [4]. The proposed rule-based similarity model makes it possible to avoid an arbitrary aggregation of marginal similarity functions. In this approach, comprehensive similarity is represented by decision rules induced from classification examples. These rules underline the monotonic relationship the more similar is object y to object x with respect to the considered features, the closer is y to x in terms of the membership to a given decision class X. Violation of this principle causes an inconsistency in the set of objects, which is handled using DRSA. An important characteristic of the proposed approach is that induced rules employ only ordinal properties of marginal similarity functions. Thus, this approach is invariant to ordinally equivalent marginal similarity functions. We improve over [4] by proposing a way o inducing decision rules, and by introducing a new rule-based classification scheme extending the one given in []. This paper is organized as follows. Section 2 describes problem setting. In Section 3, we discuss basic notions and assumptions. Section 4 defines considered similarity learning task. In Section 5, we introduce two comprehensive closeness relations. Section 6 defines rough approximations of the sets of objects being in either kind of comprehensive closeness relation with a reference object x. In Section 7, we describe induction of monotonic decision rules from the rough approximations. Section concerns application of induced rules. In Section 9, we present an illustrative example. Section 0 concludes the paper. 2 Problem Setting We consider the following classification problem setting. There is given a finite set of objects U (case base) and a finite family of pre-defined decision classes D. An object y U (a case ) is described in terms of features f,..., f n F. For each feature f i F, there is given a marginal similarity function σ fi : U U [0, ], such that the value σ fi (y, x) expresses the similarity of object y U to object x U with respect to (w.r.t.) feature f i, and for all x, y U, σ fi (y, x) = f i (y) = f i (x). Moreover, for each object y U there is given an information concerning credibility of its membership to each of the considered classes. To admit graded credibilities, each class X D is modeled as a fuzzy set in U [5], characterized by membership function µ X : U [0, ]. Thus, each object y U can belong to different decision classes with different degrees of membership. The above input information is processed to produce a recommendation concerning a new object z, in terms of a degree of membership of z to particular classes.

3 Basic Notions and Assumptions Pairwise fuzzy information base. Given the problem setting introduced in Section 2, a pairwise fuzzy information base B [5 7] is the 3-tuple B =< U, F, Σ >, () where U is a finite set of objects (a case base), F is a finite set of n features, and Σ = {σ f, σ f2,..., σ fn } is a finite set of n marginal similarity functions. Marginal similarity functions. Different marginal similarity functions can be used, depending on the value set V fi of feature f i F. For a numeric feature f i, with values on interval or ratio scale, similarity can be defined using a function, f e.g., σ fi = i(x) f i(y) max vi V fi min vi. For a nominal feature f V i, similarity can be fi defined using a table, like Table. The marginal similarity functions create an Table. Exemplary definition of similarity for a nominal feature f i F f i(x) \ f i(y) low medium high low.0 0.6 0.3 medium 0.6.0 0.5 high 0.3 0.5.0 n-dimensional similarity space. Problem decomposition. We consider the decision classes belonging to family D to be mutually independent in the sense of membership function values. Then, we decompose the original multi-class problem π to a set of single-class subproblems π X, where X D. Thus, each subproblem concerns a single decision class X D with membership function µ X : U [0, ]. In each subproblem, let V µx = {µ X (y) : y U}. (2) Reference objects. We assume that for each subproblem π X, there is given a set of so-called reference objects UX R U. These are objects to which objects from set U are going to be compared. The reference objects may be indicated by a user, and thus, the set of reference objects should be relatively small. If such information is not available, one can use clustering to choose a suitable set of reference objects, sample U, or treat all the objects from U as the reference ones. 4 Similarity Learning The method proposed in this paper is designed for the following learning task. Given: i) the pairwise fuzzy information base B, ii) the family D of decision

classes, implying subproblems π X, X D, iii) the membership functions µ X : U [0, ], X D, and iv) the sets of reference objects UX R U, X D, learn, for each subproblem π X, set of decision rules R X = R X (x), (3) x U R X where R X (x) is the set of rules describing membership of an object y U to class X D based on similarity of y to reference object x U R X. 5 Comprehensive Closeness of Objects Given a decision class X being a fuzzy set in U, we define two kinds of binary comprehensive closeness relations on U: y X α,β x µ X (x) [α, β] and µ X (y) [α, β], (4) y X α,β x µ X (x) [α, β] and µ X (y) / (α, β), (5) where y, x U and δ α β +δ, where δ R + is any fixed positive value (a technical parameter, e.g., 0.0). When y X α,β x, then α µ X(y) µ X (x) β or α µ X (x) µ X (y) β, i.e., looking from the perspective of y, µ X (y) is on the left side of µ X (x) but not farther than α, or µ X (y) is on the right side of µ X (x) but not farther than β. When y X α,β x, then µ X(y) is on the left side of µ X (x) but not closer than α, or µ X (y) is on the right side of µ X (x) but not closer than β. Thus, α and β play roles of limiting levels of membership to X. The special values δ and + δ, where δ R +, are considered in (5) to allow, respectively, µ X (y) / ( δ, β) (i.e., µ X (y) β) and µ X (y) / (α, + δ) (i.e., µ X (y) α). This is crucial, e.g., when X is crisp one can then consider two meaningful relations X 0,+δ and X δ,, composed of pairs (y, x) U U such that µ X (y) 0 and µ X (y), respectively. Let us observe that X α,β is reflexive, symmetric and transitive and thus it is an equivalence relation. Moreover, X α,β is only transitive. Given a class X and a reference object x UX R, we are interested in characterizing, in terms of similarity-based decision rules, the objects y U being in: X α,β relation with x, where α, β V µ X, X α,β relation with x, where α, β V µ X { δ} { + δ}), α < µ X (x) < β. Let Vµ δ X = V µx { δ} { + δ}, where δ R +. We define two types of sets: S( X α,β, x) = {y U : y X α,β x}, where α, β V µx, α µ X (x) β, (6) S( X α,β, x) = {y U : y X α,β x}, where α, β V δ µ X, α < µ X (x) < β. (7) The strict constraint α < µ X (x) < β in (7) prevents from considering not meaningful sets S( X α,β, x) [3]. From this point of view, it is crucial that when µ X (x) = 0 (or µ X (x) = ), one can take α = δ (or β = + δ, respectively). The sets of objects defined by (6) and (7) are to be approximated using dominance cones in the similarity space created by functions σ f,..., σ fn.

6 Rough Approximation by Dominance Relation Let us define the dominance relation w.r.t. the similarity to an object x U, called in short x-dominance relation, defined over U, and denoted by D x. For any x, y, w U, y is said to x-dominate w (denotation yd x w) if for every f i F, σ fi (y, x) σ fi (w, x). () Thus, object y is said to x-dominate object w iff for every feature f i F, y is at least as similar to x as w is. Given an object y U, x-positive and x-negative dominance cones of y in the similarity space are defined as follows: D + x (y) = {w U : wd x y}, (9) D x (y) = {w U : yd x w}. (0) In order to induce meaningful certain and possible decision rules concerning similarity to a reference object x UX R, we structure the objects y U by calculation of lower and upper approximations of sets S( X α,β, x) and S( X α,β, x). The lower approximations of sets S( X α,β, x) and S( X α,β, x) are defined as: S( X α,β, x) = {y U : D + x (y) S( X α,β, x)}, () S( X α,β, x) = {y U : D x (y) S( X α,β, x)}, (2) and the upper approximations of sets S( X α,β, x) and S( X α,β, x) are defined as: S( X α,β, x) = {y U : D x (y) S( X α,β, x) }, (3) S( X α,β, x) = {y U : D+ x (y) S( X α,β, x) }. (4) With respect to the three basic properties of set approximations defined for rough sets in [0], it follows from definitions (6), (7), (), (2), (3), and (4), that lower and upper approximations defined above fulfill properties of rough inclusion and monotonicity of the accuracy of approximation. Moreover, these approximations enjoy also complementarity property, as shown in [4]. Using (), (2), (3), and (4), one can define the boundary of set S( X α,β, x) (or set S( X α,β, x)), as the difference between its upper and lower approximation. It is also possible to perform further DRSA-like analysis by calculating the quality of approximation, reducts, and the core (see, e.g., [4,, 2]). 7 Induction of Decision Rules Lower (or upper) approximations of considered sets S( X α,β, x) and S( X α,β, x) are the basis for induction of certain (or possible) decision rules belonging to set R X (x), x UX R. We distinguish two types of rules:

() at least rules: if σ fi (y, x) h i... and σ fip (y, x) h ip, then certainly (or possibly) y X α,β x, (2) at most rules: if σ fi (y, x) h i... and σ fip (y, x) h ip, then certainly (or possibly) y X α,β x, where {f i,..., f ip } F, h i,..., h ip [0, ], and α, β satisfy 0 α µ X (x) β in case of at least rules, and δ α < µ X (x) < β + δ in case of at most rules, δ R +. Remark that according to definitions (4) and (5), the decision part of the rule of type () and (2) can be rewritten, respectively, as: () then certainly (or possibly) µ X (y) [α, β], i.e., the conclusion is that the membership of object y to decision class X is inside the interval [α, β], (2) then certainly (or possibly) µ X (y) / (α, β), i.e., the conclusion is that the membership of object y to decision class X is outside the interval (α, β). A certain rule of type () is read as: if similarity of object y to reference object x w.r.t. feature f i is at least h i... and similarity of y to x w.r.t. feature f ip is at least h ip, then certainly y belongs to class X with credibility between α and β. A possible rule of type (2) is read as: if similarity of object y to reference object x w.r.t. feature f i is at most h i... and similarity of y to x w.r.t. feature f ip is at most h ip, then possibly y belongs to class X with credibility at most α or at least β. Decision rules of type () and (2) can be induced using the VC-DomLEM algorithm [2]. On one hand, these rules reveal similarity-based patterns present in the training data. On the other hand, set R X = x U R X R X (x) of induced certain/possible rules can be applied to classify new objects (new cases). Application of Decision Rules The rules from R X can be applied to a new object z, described in terms of features f,..., f n F, to predict its degree of membership to class X. Then, the rules covering z may give an ambiguous classification suggestion (intervals of µ X instead of a crisp value). In order to resolve this ambiguity, we adapt and revise the rule classification scheme described in []. In this way, one can obtain a precise (crisp) value of membership µ X (z). Let us consider three situations, assuming that Cov z R X denotes the set of rules covering object z, Cond ρ U denotes the set of objects covered by rule ρ, UX t = {y U : µ X(y) = t}, and denotes cardinality of a set. Situation (i). No rule from R X covers object z (i.e., Cov z = ), so there is no reliable suggestion concerning µ X (z). If a concrete answer is expected, one can suggest that µ X (z) equals to the most frequent value µ X (y), where y U. Situation (ii). Exactly one rule ρ R X (x) R X, x UX R, covers object z (i.e., Cov z = ). Then, we calculate value Score ρ X (t, z) for each membership value t V µx covered by the decision part of this rule: Score ρ X (t, z) = Cond ρ U t X 2 Cond ρ U t X. (5)

Then, µ X (z) is calculated as µ X (z) = max t Score ρ X (t, z). Let us observe that Score ρ X (t, z) [0, ]. It can be interpreted as the degree of certainty of the suggestion that µ X (z) equals t. Situation (iii). Several rules from R X cover object z (i.e., Cov z > ). Then, we calculate value Score X (t, z) for each t V µx covered by the decision part of any of the covering rules: Score X (t, z) = Score + X (t, z) Score X (t, z), (6) where Score + X (t, z) and Score X (t, z) represent the positive and negative part of Score X (t, z), respectively. Score + X (t, z) takes into account rules ρ,..., ρ k Cov z whose decision part covers t: Score + X (t, z) = (Cond ρ U t X )... (Cond ρ k U t X ) 2 Cond ρ... Cond ρk U t X. (7) Let us observe that Score + X (t, z) [0, ]. Score X (t, z) takes into account the rules ρ k+,..., ρ h Cov z whose decision part does not cover t. If there is no such rule, then Score X (t, z) = 0. Otherwise: Score X (t, z) = (Cond ρ k+ U ρ k+ X )... (Cond ρ h U ρ h X ) 2 Cond ρk+... Cond ρh U ρ k+ X... U ρ h X, () where U ρ X is subset of U containing objects whose membership to class X is covered by the decision part of rule ρ. Let us observe that Score X (t, z) [0, ]. After calculating Score X (t, z) for all considered values of t, we take µ X (z) = max t Score X (t, z). It can be interpreted as a net balance of the arguments in favor and against the suggestion the membership of object z to class X equals t. 9 Illustrative Example Let us consider set U composed of five objects described by two features: f, with value set [0, ], and f 2, with value set [0, ]. Moreover, let us consider decision class X, with membership function µ X. The five objects are presented in Fig.. We assume that object x is a reference object, and that there are given two marginal similarity functions σ f, σ f2 defined as: σ fi (y, x) = f i(y) f i (x) f max f min, where i =, 2, and fi max, fi min denote max and min value in the value set of f i. Functions σ f and σ f2 create a 2-dimensional similarity space. Fig. 2 shows pairs of objects (, x) in this space. First, using (9) and (0), we calculate x-positive and x-negative dominance cones in the similarity space. Two such cones are shown in Fig. and in Fig. 2. i i

f 2 0 Dx (y ) D x + (y 3) Dx (y ) 0.6 y 2 y x 0.4 0.3 0.5 0 2 3 4 5 6 7 f y 3 y 4 0.7 Fig.. Set of objects considered in the illustrative example; the number below an object id denotes the value of function µ X for this object; the hatched area corresponds to dominance cone D x + (y 3), and the two dotted areas (one for f (y) 2, and the other for f (y) 6) correspond to dominance cone Dx (y ) σ f2 D x (y ) (y 4, x) 0.7 (y 2, x) 0.4 (y, x) 0.3 (x, x) 0.5 0 0 2 3 4 5 6 (y 3, x) 0.6 7 D + x (y 3) σ f Fig. 2. Pairs of objects (, x) in the similarity space created by σ f and σ f2 ; the number below a pair of object ids denotes the value of function µ X for the object whose id is the first in the pair; the hatched area corresponds to dominance cone D + x (y 3), and the dotted area corresponds to dominance cone D x (y ) Second, we calculate sets of objects S( X α,β, x) according to (6), for α {0.3, 0.4, 0.5} and β {0.5, 0.6, 0.7}. Moreover, we calculate S( X α,β, x) according to (7), for α { δ, 0.3, 0.4} and β {0.6, 0.7, + δ}, where δ R +. Third, sets S( X α,β, x) and S( X α,β, x) are approximated using the x-positive and x-negative dominance cones in the similarity space see Table 2. One can observe several inconsistencies w.r.t. the x-dominance relation in the similarity space. E.g., objects y 2, y 4 S( X 0.4,0.7, x) are inconsistent since they are x-dominated by object y, and y / S( X 0.4,0.7, x) (because µ X(y ) = 0.3). Table 3 presents minimal decision rules induced by VC-DomLEM algorithm from the non-empty lower and upper approximations shown in Table 2. Example of application of induced decision rules. Consider a new object z such that f (z) = 5.5, f 2 (z) = 0.5, and thus, σ f (z, x) = 6.5/, σ f2 (z, x) = /2. Object z is covered by rules ρ 2 and ρ 5, suggesting µ X (z) [0.3, 0.6] and µ X (z) / (0.4, 0.6), respectively. Applying (7), (), and (6) for each membership degree t {0.3, 0.4, 0.5, 0.6, 0.7}, we get the result shown in Table 4. Consequently, one can conclude that µ X (z) is equal to 0.3, 0.4, or 0.6.

Table 2. Approximations of sets S( X α,β, x), S( X α,β, x); δ R+; objects struck through belong to respective set but not to its lower approximation; underlined objects do not belong to respective set but belong to its upper approximation S( X α,β, x) β = 0.5 β = 0.6 β = 0.7 S( X α,β, x) β = 0.6 β = 0.7 β = + δ α = 0.3 {y, y 2, x} {y, y 2, y 3, x} U α = δ {y 3, y 4} {y 4} α = 0.4 {y 2, x} {y 2, y 3, x} {y 2, y 3, y 4, x} α = 0.3 {y, y 3, y 4} {y, y 4} {y } α = 0.5 {x} {y 3, x} {y 3, y 4, x} α = 0.4 {y, y 2, y 3, y 4} {y, y 2, y 4} {y, y 2} S( X α,β, x) β = 0.5 β = 0.6 β = 0.7 S( X α,β, x) β = 0.6 β = 0.7 β = + δ α = 0.3 {y, y 2, x} {y, y 2, y 3, x} U α = δ {y 3, y 4} {y 4} α = 0.4 {y, y 2, x} {y, y 2, y 3, x} {y, y 2, y 3, y 4, x} α = 0.3 {y, y 2, y 3, y 4} {y, y 2, y 4} {y, y 2, y 4} α = 0.5 {x} {y 3, x} {y, y 2, y 3, y 4, x} α = 0.4 {y, y 2, y 3, y 4} {y, y 2, y 4} {y, y 2, y 4} Table 3. Rules induced for referent x such that f (x) = 4, f 2(x) = 0; Supp. ( Supp. ) presents ids of objects supporting a rule (covered by a rule but not supporting it); possible rules, identical (except for possibly ) to respective certain rules, are skipped Id Decision rule Supp. Supp. ρ if σ f (y, x) 5 and σf2 (y, x), then certainly µ X(y) [0.3, 0.5] {y, y 2, x} ρ 2 if σ f (y, x) 5, then certainly µx(y) [0.3, 0.6] {y, y2, y3, x} ρ 3 if σ f (y, x), then certainly µ X(y) [0.5, 0.5] {x} ρ 4 if σ f (y, x) 7, then certainly µx(y) [0.5, 0.6] {y3, x} ρ 7 if σ f (y, x) 5 and σf2 (y, x), then possibly µ X(y) [0.4, 0.5] {y 2, x} {y } ρ if σ f (y, x) 5, then possibly µx(y) [0.4, 0.6] {y2, y3, x} {y} ρ 9 if σ f (y, x) 4, then possibly µx(y) [0.4, 0.7] {y2, y3, y4, x} {y} ρ 2 if σ f (y, x) 4, then possibly µx(y) [0.5, 0.7] {y3, y4, x} {y, y2} ρ 3 if σ f2 (y, x) 0, then certainly µ X(y) 0.6 {y 3} ρ 4 if σ f (y, x) 4, then certainly µx(y) 0.7 {y4} ρ 5 if σ f (y, x) 7, then certainly µx(y) / (0.4, 0.6) {y, y2, y3, y4} ρ 6 if σ f (y, x) 6, then certainly µx(y) / (0.4, 0.7) {y, y2, y4} ρ 9 if σ f (y, x) 7, then possibly µx(y) / (0.3, 0.6) {y, y3, y4} {y2} ρ 20 if σ f (y, x) 6, then possibly µx(y) / (0.3, 0.7) {y, y4} {y2} ρ 2 if σ f (y, x) 6, then possibly µx(y) 0.3 {y} {y2, y4} ρ 24 if σ f (y, x) 6, then possibly µx(y) 0.4 {y, y2} {y4} Table 4. Scores of a new object z resulting from application of induced decision rules t V µx 0.3 0.4 0.5 0.6 0.7 Score + X (t, z) 5 5 4 5 4 Score X (t, z) 0 0 0 Score X(t, z) 5 5 3 4 5 3 4 0 Conclusions We presented a method of similarity-based classification using the Dominancebased Rough Set Approach. This method exploits only ordinal properties of

marginal similarity functions and membership functions of decision classes. It avoids arbitrary aggregation of marginal similarities into one comprehensive similarity. Instead, it uses a rule-based similarity model employing rules describing monotonic relationship between comprehensive similarity of objects and their similarities with respect to single features. Thus, our case-based reasoning approach is as much neutral and objective as possible. Moreover, our method provides more insight when determining membership of a new object z to class X one can see the rules matching z and the objects supporting these rules. Acknowledgments. The first author wishes to acknowledge financial support from the Faculty of Computing at Poznań University of Technology, grant 09/9/DSMK/0609. References. Błaszczyński, J., Greco, S., Słowiński, R.: Multi-criteria classification a new scheme for application of dominance-based decision rules. European Journal of Operational Research (3), 030 044 (2007) 2. Błaszczyński, J., Słowiński, R., Szeląg, M.: Sequential covering rule induction algorithm for variable consistency rough set approaches. INS, 97 002 (20) 3. Gilboa, I., Schmeidler, D.: A Theory of Case-Based Decisions. No. 970520037 in Cambridge Books, Cambridge University Press (February 200) 4. Greco, S., Matarazzo, B., Słowiński, R.: Rough sets theory for multicriteria decision analysis. European Journal of Operational Research 29(), 47 (200) 5. Greco, S., Matarazzo, B., Słowiński, R.: Dominance-based rough set approach to case-based reasoning. In: Torra, V., et al. (eds.) Modelling Decisions for Artificial Intelligence. LNAI, vol. 35, pp. 7. Springer, Berlin (2006) 6. Greco, S., Matarazzo, B., Słowiński, R.: Case-based reasoning using gradual rules induced from dominance-based rough approximations. In: Wang, G., et al. (eds.) RSKT. LNAI, vol. 5009, pp. 26 275. Springer, Berlin (200) 7. Greco, S., Matarazzo, B., Słowiński, R.: Granular computing for reasoning about ordered data: the dominance-based rough set approach. In: Pedrycz, W., et al. (eds.) Handbook of Granular Computing, chap. 5, pp. 347 373. Wiley (200). Hume, D.: An Enquiry Concerning Human Understanding. Clarendon Press (74) 9. Kolodner, J.: Case-Based Reasoning. Morgan Kaufmann, San Mateo (993) 0. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning about Data. Kluwer, Dordrecht, Netherlands (99). Słowiński, R., Greco, S., Matarazzo, B.: Rough sets in decision making. In: Meyers, R. (ed.) Encyclopedia of Complexity and Systems Science, pp. 7753 776. Springer, New York (2009) 2. Słowiński, R., Greco, S., Matarazzo, B.: Rough set based decision support. In: Burke, E.K., Kendall, G. (eds.) Search Methodologies: Introductory Tutorials in Optimization and Decision Support Techniques, pp. 557 609. Springer (204) 3. Szeląg, M.: Application of the Dominance-based Rough Set Approach to Ranking and Similarity-based Classification Problems. Ph.D. thesis, Poznań University Of Technology (205), http://www.cs.put.poznan.pl/mszelag/research/mszphd.pdf 4. Szeląg, M., Greco, S., Błaszczyński, J., Słowiński, R.: Case-based reasoning using dominance-based decision rules. In: Yao, J., et al. (eds.) RSKT. LNCS, vol. 6954, pp. 404 43. Springer, Heidelberg (20) 5. Zadeh, L.: Fuzzy sets. Information and Control (3), 33 353 (965)