Machine Learning 2nd Edi7on

Similar documents
Learning Decision Trees

Learning Decision Trees

Decision Trees. Data Science: Jordan Boyd-Graber University of Maryland MARCH 11, Data Science: Jordan Boyd-Graber UMD Decision Trees 1 / 1

Lecture 3: Decision Trees

Decision Tree Learning Mitchell, Chapter 3. CptS 570 Machine Learning School of EECS Washington State University

Decision Trees.

Introduction. Decision Tree Learning. Outline. Decision Tree 9/7/2017. Decision Tree Definition

CS 6375 Machine Learning

Decision-Tree Learning. Chapter 3: Decision Tree Learning. Classification Learning. Decision Tree for PlayTennis

Supervised Learning! Algorithm Implementations! Inferring Rudimentary Rules and Decision Trees!

Decision Trees.

Decision Tree Learning

Chapter 3: Decision Tree Learning

Decision Tree Learning

Decision trees. Special Course in Computer and Information Science II. Adam Gyenge Helsinki University of Technology

Dan Roth 461C, 3401 Walnut

EECS 349:Machine Learning Bryan Pardo

Lecture 3: Decision Trees

Administration. Chapter 3: Decision Tree Learning (part 2) Measuring Entropy. Entropy Function

the tree till a class assignment is reached

Lecture 24: Other (Non-linear) Classifiers: Decision Tree Learning, Boosting, and Support Vector Classification Instructor: Prof. Ganesh Ramakrishnan

Classification and Prediction

CS6375: Machine Learning Gautam Kunapuli. Decision Trees

Imagine we ve got a set of data containing several types, or classes. E.g. information about customers, and class=whether or not they buy anything.

Question of the Day. Machine Learning 2D1431. Decision Tree for PlayTennis. Outline. Lecture 4: Decision Tree Learning

Decision Trees. Tirgul 5

Decision Tree Learning and Inductive Inference

Learning Classification Trees. Sargur Srihari

Decision Trees. Gavin Brown

Decision Trees / NLP Introduction

Outline. Training Examples for EnjoySport. 2 lecture slides for textbook Machine Learning, c Tom M. Mitchell, McGraw Hill, 1997

Machine Learning 2nd Edition

Classification: Decision Trees

Machine Learning & Data Mining

Introduction to ML. Two examples of Learners: Naïve Bayesian Classifiers Decision Trees

Decision Trees. CS57300 Data Mining Fall Instructor: Bruno Ribeiro

DECISION TREE LEARNING. [read Chapter 3] [recommended exercises 3.1, 3.4]

Chapter 3: Decision Tree Learning (part 2)

M chi h n i e n L e L arni n n i g Decision Trees Mac a h c i h n i e n e L e L a e r a ni n ng


Decision Tree Analysis for Classification Problems. Entscheidungsunterstützungssysteme SS 18

Classification: Decision Trees

Decision Trees Part 1. Rao Vemuri University of California, Davis

Classification and regression trees

2018 CS420, Machine Learning, Lecture 5. Tree Models. Weinan Zhang Shanghai Jiao Tong University

Decision Tree Learning - ID3

Introduction to Machine Learning CMU-10701

C4.5 - pruning decision trees

Decision Tree Learning

Decision Trees. Danushka Bollegala

Classification II: Decision Trees and SVMs

CS145: INTRODUCTION TO DATA MINING

Notes on Machine Learning for and

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Intelligent Data Analysis. Decision Trees

Symbolic methods in TC: Decision Trees

Artificial Intelligence. Topic

Machine Learning Recitation 8 Oct 21, Oznur Tastan

Data Mining. CS57300 Purdue University. Bruno Ribeiro. February 8, 2018

Decision Trees. Each internal node : an attribute Branch: Outcome of the test Leaf node or terminal node: class label.

Einführung in Web- und Data-Science

Machine Learning: Symbolische Ansätze. Decision-Tree Learning. Introduction C4.5 ID3. Regression and Model Trees

Typical Supervised Learning Problem Setting

Machine Learning 3. week

Introduction Association Rule Mining Decision Trees Summary. SMLO12: Data Mining. Statistical Machine Learning Overview.

Rule Generation using Decision Trees

Symbolic methods in TC: Decision Trees

Apprentissage automatique et fouille de données (part 2)

Decision Trees. Lewis Fishgold. (Material in these slides adapted from Ray Mooney's slides on Decision Trees)

Lecture 7 Decision Tree Classifier

Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation

Decision Trees. CSC411/2515: Machine Learning and Data Mining, Winter 2018 Luke Zettlemoyer, Carlos Guestrin, and Andrew Moore

Decision Tree. Decision Tree Learning. c4.5. Example

CS 380: ARTIFICIAL INTELLIGENCE MACHINE LEARNING. Santiago Ontañón

Classification Using Decision Trees

Decision Tree And Random Forest

Decision Trees Entropy, Information Gain, Gain Ratio

Inductive Learning. Chapter 18. Material adopted from Yun Peng, Chuck Dyer, Gregory Piatetsky-Shapiro & Gary Parker

10-701/ Machine Learning: Assignment 1

Decision Tree Learning Lecture 2

Classification and Regression Trees

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction

ML techniques. symbolic techniques different types of representation value attribute representation representation of the first order

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag

UVA CS 4501: Machine Learning

Decision Tree Learning

Lecture 7: DecisionTrees

Chapter 6: Classification

ARTIFICIAL INTELLIGENCE. Supervised learning: classification

Artificial Intelligence Decision Trees

The Naïve Bayes Classifier. Machine Learning Fall 2017

Algorithms for Classification: The Basic Methods

Decision Tree Learning

The Quadratic Entropy Approach to Implement the Id3 Decision Tree Algorithm

CSCI 5622 Machine Learning

Decision Tree Learning

Decision Support. Dr. Johan Hagelbäck.

The Solution to Assignment 6

Tutorial 6. By:Aashmeet Kalra

Ensemble Methods. Charles Sutton Data Mining and Exploration Spring Friday, 27 January 12

Induction on Decision Trees

Transcription:

Lecture Slides for INTRODUCTION TO Machine Learning 2nd Edi7on CHAPTER 9: Decision Trees ETHEM ALPAYDIN The MIT Press, 2010 Edited and expanded for CS 4641 by Chris Simpkins alpaydin@boun.edu.tr h1p://www.cmpe.boun.edu.tr/~ethem/i2ml2e

Overview Univariate decision trees Building classification trees Dealing with overfitting Extracting rules from decision trees 2

Tree Uses Nodes, and Leaves 3

Divide and Conquer Internal decision nodes Univariate: Uses a single atribute, x i Numeric x i : Binary split : x i > w m Discrete x i : n- way split for n possible values Mul7variate: Uses all atributes, x Leaves Classifica7on: Class labels, or propor7ons Regression: Numeric; r average, or local fit Learning is greedy; find the best split recursively (Breiman et al, 1984; Quinlan, 1986, 1993) 4

Classifica7on Trees (ID3, CART, C4.5) For node m, N m instances reach m, N i m belong to C i Node m is pure if p i m is 0 or 1 Measure of impurity is entropy 5

Best Split If node m is pure, generate a leaf and stop, otherwise split and con7nue recursively Impurity aber split: N mj of N m take branch j. N i mj belong to C i Find the variable and split that min impurity (among all variables - - and split posi7ons for numeric variables) 6

7

Summary of Main Loop Pick A, the best decision attribute for examples at current node using impurity measure For each value of A, create new descendant leaf nodes of current node Sort training examples to leaf nodes according to their values for attribute A If training examples perfectly classified, Then STOP, Else iterate over new leaf nodes (turning some of them into internal nodes) 8

Example: Play tennis today? 1 Day Outlook Temperatur Humidity Wind PlayTennis? D1 Sunny ehot High Weak No D2 Sunny Hot High Strong No D3 Overcast Hot High Weak Yes D4 Rain Mild High Weak Yes D5 Rain Cool Normal Weak Yes D6 Rain Cool Normal Strong No D7 Overcast Cool Normal Strong Yes D8 Sunny Mild High Weak No D9 Sunny Cool Normal Weak Yes D10 Rain Mild Normal Weak Yes D11 Sunny Mild Normal Strong Yes D12 Overcast Mild High Strong Yes D13 Overcast Hot Normal Weak Yes D14 Rain Mild High Strong No 1 Example from Mitchell 1997 9

Choosing the first split atribute Outlook: Temperature: Humidity: Wind: 10

11

Choosing an atribute for Sunny node X Sunny = {D1, D2, D8, D9, D11} Day Outlook Temperatur Humidity Wind PlayTennis? D1 Sunny ehot High Weak No D2 Sunny Hot High Strong No D8 Sunny Mild High Weak No D9 Sunny Cool Normal Weak Yes D11 Sunny Mild Normal Strong Yes Humidity: Do same calculation for Temperature and Wind Humidity has lowest entropy, so Humidity is split attribute at Sunny node 12

Final Induced Tree 13

Observa7ons Temperature was not needed Decision tree can be used for feature extraction (a.k.a. feature selection) Tree is short ID3 family of algorithms have preference bias for short trees Occam s Razor: shorter trees are better 14

Pruning Trees Remove subtrees for beter generaliza7on (decrease variance) Prepruning: Early stopping Postpruning: Grow the whole tree then prune subtrees which overfit on the pruning set Prepruning is faster, postpruning is more accurate (requires a separate pruning set) 15

Rule Extrac7on from Trees C4.5Rules (Quinlan, 1993) to Machine Learning The MIT Press (V1.1) 16

Conclusion Decision trees good when: Instances described by attribute- value pairs Target function is discrete Interpretability of learned hypothesis (e.g., as a rule set) is desired Training data may be noisy Decision trees are a good first algorithm to try to Machine Learning The MIT Press (V1.1) 17