Outline. [read Chapter 2] Learning from examples. General-to-specic ordering over hypotheses. Version spaces and candidate elimination.

Similar documents
Concept Learning through General-to-Specific Ordering

[read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] General-to-specific ordering over hypotheses

Fundamentals of Concept Learning

Version Spaces.

Outline. Training Examples for EnjoySport. 2 lecture slides for textbook Machine Learning, c Tom M. Mitchell, McGraw Hill, 1997

Concept Learning.

Concept Learning Mitchell, Chapter 2. CptS 570 Machine Learning School of EECS Washington State University

Question of the Day? Machine Learning 2D1431. Training Examples for Concept Enjoy Sport. Outline. Lecture 3: Concept Learning

Overview. Machine Learning, Chapter 2: Concept Learning

Machine Learning 2D1431. Lecture 3: Concept Learning

Concept Learning. Space of Versions of Concepts Learned

CSCE 478/878 Lecture 2: Concept Learning and the General-to-Specific Ordering

Lecture 2: Foundations of Concept Learning

Introduction to machine learning. Concept learning. Design of a learning system. Designing a learning system

Concept Learning. Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University.

EECS 349: Machine Learning

Concept Learning. Berlin Chen References: 1. Tom M. Mitchell, Machine Learning, Chapter 2 2. Tom M. Mitchell s teaching materials.

Computational Learning Theory

Answers Machine Learning Exercises 2

Computational Learning Theory

EECS 349: Machine Learning Bryan Pardo

Computational Learning Theory

BITS F464: MACHINE LEARNING

Topics. Concept Learning. Concept Learning Task. Concept Descriptions

Machine Learning. Computational Learning Theory. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Computational Learning Theory. Eric Xing , Fall Lecture 9, October 5, 2016

Computational Learning Theory

Introduction to Machine Learning

Introduction to Machine Learning

[Read Ch. 5] [Recommended exercises: 5.2, 5.3, 5.4]

Notes on Machine Learning for and

Linear Classifiers: Expressiveness

Lecture 3: Decision Trees

Data Mining and Machine Learning

Classification (Categorization) CS 391L: Machine Learning: Inductive Classification. Raymond J. Mooney. Sample Category Learning Problem

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Lecture 3: Decision Trees

Tutorial 6. By:Aashmeet Kalra

Bayesian Learning Features of Bayesian learning methods:

Decision Tree Learning and Inductive Inference

Learning Decision Trees

Computational Learning Theory

Machine Learning 2007: Slides 1. Instructor: Tim van Erven Website: erven/teaching/0708/ml/

EECS 349:Machine Learning Bryan Pardo

CSCE 478/878 Lecture 6: Bayesian Learning

Decision Tree Learning

The Bayesian Learning

Decision Trees.

Decision Trees.

Chapter 3: Decision Tree Learning

Decision Tree Learning

Lecture 9: Bayesian Learning

Machine Learning: Exercise Sheet 2

Decision Trees. Danushka Bollegala

A Tutorial on Computational Learning Theory Presented at Genetic Programming 1997 Stanford University, July 1997

T Machine Learning: Basic Principles

Lecture 25 of 42. PAC Learning, VC Dimension, and Mistake Bounds

Introduction to ML. Two examples of Learners: Naïve Bayesian Classifiers Decision Trees

Question of the Day. Machine Learning 2D1431. Decision Tree for PlayTennis. Outline. Lecture 4: Decision Tree Learning

CS 6375 Machine Learning

Introduction to Machine Learning

Learning Sets of Rules. Sequential covering algorithms FOIL. Induction as inverse of deduction. Inductive Logic Programming

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag

B555 - Machine Learning - Homework 4. Enrique Areyan April 28, 2015

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction

Learning Decision Trees

Bayesian Classification. Bayesian Classification: Why?

Artificial Intelligence. Topic

Learning Classification Trees. Sargur Srihari

Decision Tree Learning Mitchell, Chapter 3. CptS 570 Machine Learning School of EECS Washington State University

Machine Learning

Decision Tree Learning

Decision-Tree Learning. Chapter 3: Decision Tree Learning. Classification Learning. Decision Tree for PlayTennis

Introduction. Decision Tree Learning. Outline. Decision Tree 9/7/2017. Decision Tree Definition

Arguing and Explaining Classifications

Algorithms for Classification: The Basic Methods

Relationship between Least Squares Approximation and Maximum Likelihood Hypotheses

Machine Learning

Course 395: Machine Learning

Data classification (II)

Machine Learning Recitation 8 Oct 21, Oznur Tastan

Computational learning theory. PAC learning. VC dimension.

Overview. I Review of natural deduction. I Soundness and completeness. I Semantics of propositional formulas. I Soundness proof. I Completeness proof.

CS206 Lecture 03. Propositional Logic Proofs. Plan for Lecture 03. Axioms. Normal Forms

Lecture Notes in Machine Learning Chapter 4: Version space learning

Supervised Learning! Algorithm Implementations! Inferring Rudimentary Rules and Decision Trees!

Applied Logic. Lecture 4 part 2 Bayesian inductive reasoning. Marcin Szczuka. Institute of Informatics, The University of Warsaw

Hypothesis Evaluation

MODULE -4 BAYEIAN LEARNING

Computational Learning Theory (COLT)

Topics. Bayesian Learning. What is Bayesian Learning? Objectives for Bayesian Learning

Typical Supervised Learning Problem Setting

Logic: Bottom-up & Top-down proof procedures

Individual learning and population evolution

Mining Classification Knowledge

Machine Learning 2010

Decision Trees. Tirgul 5

Stephen Scott.

The Naïve Bayes Classifier. Machine Learning Fall 2017

Mining Classification Knowledge

Transcription:

Outline [read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] Learning from examples General-to-specic ordering over hypotheses Version spaces and candidate elimination algorithm Picking new examples The need for inductive bias Note: simple approach assuming no noise, illustrates key concepts 22 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Training Examples for EnjoySport Sky Temp Humid Wind Water Forecst EnjoySpt Sunny Warm Normal Strong Warm Same Yes Sunny Warm High Strong Warm Same Yes Rainy Cold High Strong Warm Change No Sunny Warm High Strong Cool Change Yes What is the general concept? 23 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Representing Hypotheses Many possible representations Here, h is conjunction of constraints on attributes Each constraint can be aspeccvalue (e.g., Water = Warm) don't care (e.g., \W ater =?") no value allowed (e.g.,\water= ") For example, Sky AirTemp Humid Wind Water Forecst hsunny?? Strong? Samei 24 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Prototypical Concept Learning Task Given: { Instances X: Possible days, each described by the attributes Sky, AirTemp, Humidity, Wind, Water, Forecast { Target function c: EnjoySport : X!f0 1g { Hypotheses H: Conjunctions of literals. E.g. h? Cold High???i: { Training examples D: Positive and negative examples of the target function hx 1 c(x 1 )i :::hx m c(x m )i Determine: A hypothesis h in H such that h(x) =c(x) for all x in D. 25 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

The inductive learning hypothesis: Any hypothesis found to approximate the target function well over a suciently large set of training examples will also approximate the target function well over other unobserved examples. 26 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Instance, Hypotheses, and More- General-Than Instances X Hypotheses H Specific x 1 x 2 h 1 h 2 h 3 General x 1 = <Sunny, Warm, High, Strong, Cool, Same> x = <Sunny, Warm, High, Light, Warm, Same> 2 h 1 = <Sunny,?,?, Strong,?,?> h = <Sunny,?,?,?,?,?> 2 h = <Sunny,?,?,?, Cool,?> 3 27 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Find-S Algorithm 1. Initialize h to the most specic hypothesis in H 2. For each positive training instance x For each attribute constraint a i in h If the constraint a i in h is satised by x Then do nothing Else replace a i in h by the next more general constraint that is satised by x 3. Output hypothesis h 28 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Hypothesis Space Search by Find-S Instances X Hypotheses H - x 3 h 0 h 1 Specific x + 1 x+ 2 h 2,3 x+ 4 h 4 General x = <Sunny Warm Normal Strong Warm Same>, + 1 x 2 = <Sunny Warm High Strong Warm Same>, + x 3 = <Rainy Cold High Strong Warm Change>, - x = <Sunny Warm High Strong Cool Change>, + 4 h = <,,,,, > 0 h 1 = <Sunny Warm Normal Strong Warm Sam h 2 = <Sunny Warm? Strong Warm Same> h = <Sunny Warm? Strong Warm Same> 3 h = <Sunny Warm? Strong?? > 4 29 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Complaints about Find-S Can't tell whether it has learned concept Can't tell when training data inconsistent Picks a maximally specic h (why?) Depending on H, there might beseveral! 30 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Version Spaces Ahypothesis h is consistent with a set of training examples D of target concept c if and only if h(x) = c(x) for each training example hx c(x)i in D. Consistent(h D) (8hx c(x)i 2D) h(x) =c(x) The version space, VS H D, with respect to hypothesis space H and training examples D, is the subset of hypotheses from H consistent with all training examples in D. VS H D fh 2 HjConsistent(h D)g 31 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

The List-Then-Eliminate Algorithm: 1. V ersionspace a list containing every hypothesis in H 2. For each training example, hx c(x)i remove from V ersionspace any hypothesis h for which h(x) 6= c(x) 3. Output the list of hypotheses in V ersionspace 32 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Example Version Space S: { <Sunny, Warm,?, Strong,?,?> } <Sunny,?,?, Strong,?,?> <Sunny, Warm,?,?,?,?> <?, Warm,?, Strong,?,?> G: { <Sunny,?,?,?,?,?>, <?, Warm,?,?,?,?> } 33 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Representing Version Spaces The General boundary, G,of version space VS H D is the set of its maximally general members The Specic boundary, S,of version space VS H D is the set of its maximally specic members Every member of the version space lies between these boundaries VS H D = fh 2 Hj(9s 2 S)(9g 2 G)(g h s)g where x y means x is more general or equal to y 34 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Candidate Elimination Algorithm G maximally general hypotheses in H S maximally specic hypotheses in H For each training example d, do If d is a positive example { Remove from G any hypothesis inconsistent with d { For each hypothesis s in S that is not consistent with d Remove s from S Add to S all minimal generalizations h of s such that 1. h is consistent with d, and 2. some member of G is more general than h Remove from S any hypothesis that is more general than another hypothesis in S If d is a negative example 35 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

{ Remove from S any hypothesis inconsistent with d { For each hypothesis g in G that is not consistent with d Remove g from G Add to G all minimal specializations h of g such that 1. h is consistent with d, and 2. some member of S is more specic than h Remove from G any hypothesis that is less general than another hypothesis in G 36 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Example Trace S 0 : {<Ø, Ø, Ø, Ø, Ø, Ø>} G 0 : {<?,?,?,?,?,?>} 37 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

What Next Training Example? S: { <Sunny, Warm,?, Strong,?,?> } <Sunny,?,?, Strong,?,?> <Sunny, Warm,?,?,?,?> <?, Warm,?, Strong,?,?> G: { <Sunny,?,?,?,?,?>, <?, Warm,?,?,?,?> } 38 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

How Should These Be Classied? S: { <Sunny, Warm,?, Strong,?,?> } <Sunny,?,?, Strong,?,?> <Sunny, Warm,?,?,?,?> <?, Warm,?, Strong,?,?> G: { <Sunny,?,?,?,?,?>, <?, Warm,?,?,?,?> } hsunny W arm Normal Strong Cool Changei hrainy Cool Normal Light W arm Samei hsunny W arm Normal Light W arm Samei 39 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

What Justies this Inductive Leap? + hsunny W arm Normal Strong Cool Changei + hsunny W arm Normal Light W arm Samei S : hsunny W arm Normal???i Why believe we can classify the unseen hsunny W arm Normal Strong W arm Samei 40 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

An UNBiased Learner Idea: Choose H that expresses every teachable concept (i.e., H is the power set of X) Consider H = disjunctions, conjunctions, 0 negations over previous H. E.g., hsunny W arm Normal???i_:h?????Changei What are S, G in this case? S G 41 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Inductive Bias Consider concept learning algorithm L instances X, target concept c training examples D c = fhx c(x)ig let L(x i D c ) denote the classication assigned to the instance x i by L after training on data D c. Denition: The inductive bias of L is any minimal set of assertions B such that for any target concept c and corresponding training examples D c (8x i 2 X)[(B ^ D c ^ x i ) ` L(x i D c )] where A ` B means A logically entails B 42 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Inductive Systems and Equivalent Deductive Systems Training examples New instance Inductive system Candidate Elimination Algorithm Using Hypothesis Space H Classification of new instance, or "don t know" Training examples New instance Assertion " H contains the target concept" Equivalent deductive system Theorem Prover Classification of new instance, or "don t know" Inductive bias made explicit 43 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Three Learners with Dierent Biases 1. Rote learner: Store examples, Classify x i it matches previously observed example. 2. Version space candidate elimination algorithm 3. Find-S 44 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997

Summary Points 1. Concept learning as search through H 2. General-to-specic ordering over H 3. Version space candidate elimination algorithm 4. S and G boundaries characterize learner's uncertainty 5. Learner can generate useful queries 6. Inductive leaps possible only if learner is biased 7. Inductive learners can be modelled by equivalent deductive systems 45 lecture slides for textbook Machine Learning, T. Mitchell, McGraw Hill, 1997