Dariusz Brzeziński, Zbigniew Grudziński, Izabela Szczęch. Poznan University of Technology, Poland

Similar documents
Relationship between Loss Functions and Confirmation Measures

Introduction to Machine Learning

Introduction to Machine Learning

Parts 3-6 are EXAMPLES for cse634

Interpreting Low and High Order Rules: A Granular Computing Approach

A new Approach to Drawing Conclusions from Data A Rough Set Perspective

Information Theory & Decision Trees

CS 6375 Machine Learning

Machine Learning 2010

ECE521 Lecture7. Logistic Regression

The Foundations of Causal Decision Theory, by James Joyce, Cambridge: Cambridge University Press, 1999, Pp. xii H/b

Pattern-Based Decision Tree Construction

Decision Rule Approach

Reasoning Under Uncertainty

Foundations of Classification

Machine Learning & Data Mining

EECS 349:Machine Learning Bryan Pardo

Lecture VII: Classification I. Dr. Ouiem Bchir

Course 395: Machine Learning

Data Mining. CS57300 Purdue University. Bruno Ribeiro. February 8, 2018

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts

REASONING UNDER UNCERTAINTY: CERTAINTY THEORY

Applied Logic. Lecture 4 part 2 Bayesian inductive reasoning. Marcin Szczuka. Institute of Informatics, The University of Warsaw

Classification Based on Logical Concept Analysis

Fuzzy Systems. Introduction

Tutorial on Venn-ABERS prediction

Reasoning with Uncertainty

Naïve Bayesian. From Han Kamber Pei

Ensemble Methods. NLP ML Web! Fall 2013! Andrew Rosenberg! TA/Grader: David Guy Brizan

Probability, Confirmation, and the Conjunction Fallacy

A PRIMER ON ROUGH SETS:

Pointwise Exact Bootstrap Distributions of Cost Curves

Knowledge Extraction from DBNs for Images

Lecture 3: Decision Trees

Roman Słowiński. Rough or/and Fuzzy Handling of Uncertainty?

Philosophy 148 Announcements & Such

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

Jeffrey D. Ullman Stanford University

CS 188: Artificial Intelligence. Outline

I. Induction, Probability and Confirmation: Introduction

Remarks on Random Sequences

Evaluation. Andrea Passerini Machine Learning. Evaluation

Rough sets, decision algorithms and Bayes' theorem

IMBALANCED DATA. Phishing. Admin 9/30/13. Assignment 3: - how did it go? - do the experiments help? Assignment 4. Course feedback

Notes on Machine Learning for and

Topic 3: Introduction to Statistics. Algebra 1. Collecting Data. Table of Contents. Categorical or Quantitative? What is the Study of Statistics?!

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction

Evaluation requires to define performance measures to be optimized

Machine Learning 3. week

Decision Trees. Introduction. Some facts about decision trees: They represent data-classification models.

Ensemble learning 11/19/13. The wisdom of the crowds. Chapter 11. Ensemble methods. Ensemble methods

Favoring, Likelihoodism, and Bayesianism

Introduction to Machine Learning Midterm Exam

Class 4: Classification. Quaid Morris February 11 th, 2011 ML4Bio

the tree till a class assignment is reached

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Artificial Intelligence. Propositional logic

A first model of learning

Performance Evaluation and Hypothesis Testing

Machine Learning Linear Classification. Prof. Matteo Matteucci

Informal Definition: Telling things apart

Remarks on Random Sequences

Stephen Scott.

From statistics to data science. BAE 815 (Fall 2017) Dr. Zifei Liu

Confirmation Theory. Pittsburgh Summer Program 1. Center for the Philosophy of Science, University of Pittsburgh July 7, 2017

Solving Classification Problems By Knowledge Sets

A Posteriori Corrections to Classification Methods.

CSE 473: Artificial Intelligence Autumn Topics

Lecture Notes in Machine Learning Chapter 4: Version space learning

Fuzzy Systems. Introduction

THE LOGIC OF COMPOUND STATEMENTS

Weighting Expert Opinions in Group Decision Making for the Influential Effects between Variables in a Bayesian Network Model

Computational Learning Theory. Definitions

Supervised Learning! Algorithm Implementations! Inferring Rudimentary Rules and Decision Trees!

Similarity-based Classification with Dominance-based Decision Rules

MLCC 2017 Regularization Networks I: Linear Models

Feature Selection with Fuzzy Decision Reducts

CS 188: Artificial Intelligence Fall 2008

General Naïve Bayes. CS 188: Artificial Intelligence Fall Example: Overfitting. Example: OCR. Example: Spam Filtering. Example: Spam Filtering

Calcul de motifs sous contraintes pour la classification supervisée

Lecture 3: Decision Trees

Machine Learning and Deep Learning! Vincent Lepetit!

Human interpretation and reasoning about conditionals

Statistically Significant Dependencies for Spatial Co-location Pattern Mining and Classification Association Rule Discovery.

Chapter 7 Towards a Grammar of Bayesian Confirmation

PROBABILITY: CONDITIONING, BAYES THEOREM [DEVORE 2.4]

1 The Probably Approximately Correct (PAC) Model

Multiclass Classification-1

Generalization, Overfitting, and Model Selection

COMP219: Artificial Intelligence. Lecture 19: Logic for KR

Discrete Multivariate Statistics

Statistical Machine Learning from Data

Data mining, 4 cu Lecture 5:

Mapping Vulnerability and Risk of Mangrove Conversion to Pond Aquaculture in Myanmar

In Defense of Jeffrey Conditionalization

Toward a Quantum Theory of Cognition: History, Development and Perspectives

Data Mining and Machine Learning (Machine Learning: Symbolische Ansätze)

Tackling the Poor Assumptions of Naive Bayes Text Classifiers

COMP9444: Neural Networks. Vapnik Chervonenkis Dimension, PAC Learning and Structural Risk Minimization

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Transcription:

Dariusz Brzeziński, Zbigniew Grudziński, Izabela Szczęch Poznan University of Technology, Poland

Rule-based classification Confirmation measures CM-CAR algorithm Experiments Conclusions Bayesian Confirmation Measures in Rule-based Classification 2

One the most popular classification models when working with human experts Consequence relation: if E then H E H Evidence Hypothesis In classification rules H is a class label Ex: talk=short and slides=funny audience=happy Bayesian Confirmation Measures in Rule-based Classification 3

Used to find associations rather than predict Same E H relation, but H can be any attribute Usually (too) many rules are found Common task: filter out only the most interesting rules Rule Rule Rule Bayesian Confirmation Measures in Rule-based Classification 4

Height Hair Eyes Nationality tall blond blue Swede medium dark hazel German medium blond blue Swede tall blond blue German short red blue German medium dark hazel Swede E H E H E H E H E H E H H H H H E a c E 1 0 E b d E 2 3 if (Hair = red) & (Eyes = blue) then (Nationality = German) if Evidence then Hypothesis The contingency table is a form used to calculate the value of interestingness measures sup E H = a conf E H = a a + c H H E a c a + c E b d b + d a + b c + d n 5

Measures that satisfy 0 c ( H,E ) 0 0 if if if a/ P ( Ha Ec ) a/ P ( Ha Ec ) a/ P ( Ha Ec ) Confirmation measures say what is a value of information that E adds to the credibility of H Intuition: evidence should support the hypothesis Ex: talk=long and slides=boring ( Pa ( H b), )/n, ( Pa ( H b), )/n, ( Pa ( H b). )/n. Good Neutral Bad audience=happy confidence > 0, confirmation definitely < 0 Bayesian Confirmation Measures in Rule-based Classification 6

H H E a c a + c E b d b + d a + b c + d n The values of the presented measures range from -1 to 1 Bayesian Confirmation Measures in Rule-based Classification 7

1. Can confirmation measures be applied to predictive classification problems? 2. How to discover and prune decision rules with high confirmation? 3. Which confirmation measures are best suited for classification?? Bayesian Confirmation Measures in Rule-based Classification 8

Algorithm for creating classification association rules Generalization of CBA algorithm Tries to create predictive and descriptive rule lists Main idea Use two seperate sets of (confirmation) measures to select and sort classification association rules Rule Y Rule Z Rule X Rule Z > Rule X Pruning Sorting Bayesian Confirmation Measures in Rule-based Classification 9

12 confirmation measures 20 datasets: 10 balanced, 10 imbalanced ~10,000 rules generated per dataset 1%-100% rules left after pruning Comparison of accuracy, AUC, F1-score, and G-mean CM-CAR: Confirmation measure used only for rule list pruning Confirmation measure used for rule sorting and pruning Bayesian Confirmation Measures in Rule-based Classification 10

Bayesian Confirmation Measures in Rule-based Classification 11

Bayesian Confirmation Measures in Rule-based Classification 12

Critical distance plot for rule pruning Critical distance plot for rule sorting Bayesian Confirmation Measures in Rule-based Classification 13

Confirmation measures influenced the predictive performance of decision rule lists Slightly different results for rule sorting and pruning To achieve good performance on imbalanced data coverage should be additionally controlled F, Z, c 1, S performed better/comparable to the baseline Full results for accuracy, AUC, F1-score, and G-mean: http://www.cs.put.poznan.pl/dbrzezinski/software/cmcar.html Bayesian Confirmation Measures in Rule-based Classification 14

CM-CAR: algorithm for sorting and pruning rule lists based on any interestingness measure The 12 analyzed measures differed in terms of resulting classifier performance Measures F, Z, c 1, S comparable or better than conf in terms of rule sorting and pruning Future work: algorithms using confirmation measures during rule generation Bayesian Confirmation Measures in Rule-based Classification 15

Thank you! Bayesian Confirmation Measures in Rule-based Classification 16

Bayesian Confirmation Measures in Rule-based Classification 17

Desirable property of c(h,e) = f(a,b,c,d) : monotonicity (M)* f should be non-decreasing with respect to a and d and non-increasing with respect to b and c Interpretation of (M): (EH if x is a raven, then x is black) a) the more black ravens we observe, the more credible becomes EH b) the more black non-ravens we observe, the less credible becomes EH c) the more non-black ravens we observe, the less credible becomes EH d) the more non-black non-ravens we observe, the more credible becomes EH *S.Greco, Z.Pawlak, R.Słowiński: Can Bayesian confirmation measures be useful for rough set decision rules? Engineering Applications of Artificial Intelligence, 17 (2004) no.4, 345-361 18

Desirable property of c(h,e): maximality/minimality* c(h,e) is maximal if and only if P(E, H) = P( E,H) = 0 and c(h,e) is minimal if and only if P(E,H) = P( E, H) = 0. Interpretation of maximality/minimality: a measure obtains its maximum iff c=b=0 and its minimum iff a=d=0. *Glass, D.H.: Confirmation measures of association rule interestingness, Knowledge-Based Systems 44, (2013) 65 77 19

Desirable property of c(h,e): hypothesis symmetry (HS)* c(h,e) = c( H,E) Interpretation of (HS): (EH if x is a square, then x is rectangle) the strength with which the premise (x is a square) confirms the conclusion (x is rectangle) is the same as the strength with which the premise disconfirms the negated conclsuion (x is not a rectangle). *Carnap, R.: Logical Foundations of Probability, second ed. University of Chicago Press, Chicago (1962) Eells, E., Fitelson, B.: Symmetries and asymmetries in evidential support. Philosophical Studies, 107 (2) (2002), 129-142 20