Decision Tree Learning

Slides:



Advertisements
Similar presentations
1 Machine Learning: Lecture 3 Decision Tree Learning (Based on Chapter 3 of Mitchell T.., Machine Learning, 1997)
Advertisements

Decision Tree Learning
Decision Tree Learning - ID3
Decision Trees Decision tree representation ID3 learning algorithm
Machine Learning III Decision Tree Induction
1er. Escuela Red ProTIC - Tandil, de Abril, Decision Tree Learning 3.1 Introduction –Method for approximation of discrete-valued target functions.
Decision Tree Algorithm (C4.5)
ICS320-Foundations of Adaptive and Learning Systems
Classification Techniques: Decision Tree Learning
Decision Tree Example MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way.
ID3 Algorithm Abbas Rizvi CS157 B Spring What is the ID3 algorithm? ID3 stands for Iterative Dichotomiser 3 Algorithm used to generate a decision.
Machine Learning II Decision Tree Induction CSE 473.
Decision Tree Learning
Part 7.3 Decision Trees Decision tree representation ID3 learning algorithm Entropy, information gain Overfitting.
Lazy Associative Classification By Adriano Veloso,Wagner Meira Jr., Mohammad J. Zaki Presented by: Fariba Mahdavifard Department of Computing Science University.
Decision Tree Learning Learning Decision Trees (Mitchell 1997, Russell & Norvig 2003) –Decision tree induction is a simple but powerful learning paradigm.
Induction of Decision Trees
Classification and Prediction by Yen-Hsien Lee Department of Information Management College of Management National Sun Yat-Sen University March 4, 2003.
Machine Learning Lecture 10 Decision Trees G53MLE Machine Learning Dr Guoping Qiu1.
Decision tree learning
Machine Learning Chapter 3. Decision Tree Learning
Artificial Intelligence 7. Decision trees
Machine Learning Decision Tree.
Mohammad Ali Keyvanrad
Decision tree learning Maria Simi, 2010/2011 Inductive inference with decision trees  Decision Trees is one of the most widely used and practical methods.
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
Machine Learning Lecture 10 Decision Tree Learning 1.
CpSc 810: Machine Learning Decision Tree Learning.
Decision-Tree Induction & Decision-Rule Induction
Decision Tree Learning
A C B. Will I play tennis today? Features – Outlook: {Sun, Overcast, Rain} – Temperature:{Hot, Mild, Cool} – Humidity:{High, Normal, Low} – Wind:{Strong,
Elsayed Hemayed Data Mining Course
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
Iterative Dichotomiser 3 By Christopher Archibald.
CSE343/543 Machine Learning: Lecture 4.  Chapter 3: Decision Trees  Weekly assignment:  There are lot of applications and systems using machine learning.
Review of Decision Tree Learning Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
CSE573 Autumn /09/98 Machine Learning Administrative –Last topic: Decision Tree Learning Reading: 5.1, 5.4 Last time –finished NLP sample system’s.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
ICS320-Foundations of Adaptive and Learning Systems
Machine Learning Inductive Learning and Decision Trees
DECISION TREES An internal node represents a test on an attribute.
Università di Milano-Bicocca Laurea Magistrale in Informatica
CS 9633 Machine Learning Decision Tree Learning
Decision Tree Learning
Decision trees (concept learnig)
Machine Learning Lecture 2: Decision Tree Learning.
Decision trees (concept learnig)
Classification Algorithms
Decision Tree Learning
CSE543: Machine Learning Lecture 2: August 6, 2014
Teori Keputusan (Decision Theory)
Decision Trees: Another Example
Introduction to Machine Learning Algorithms in Bioinformatics: Part II
Decision Tree Saed Sayad 9/21/2018.
ID3 Algorithm.
Chapter 8 Tutorial.
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning: Lecture 3
Decision Trees Decision tree representation ID3 learning algorithm
Play Tennis ????? Day Outlook Temperature Humidity Wind PlayTennis
Machine Learning Chapter 3. Decision Tree Learning
Decision Trees.
COMP61011 : Machine Learning Decision Trees
Decision Trees Decision tree representation ID3 learning algorithm
Data Classification for Data Mining
Artificial Intelligence 6. Decision Tree Learning
INTRODUCTION TO Machine Learning
Machine Learning: Decision Tree Learning
Data Mining CSCI 307, Spring 2019 Lecture 15
Decision Tree.
Presentation transcript:

Decision Tree Learning Learning Decision Trees Decision tree induction is a simple but powerful learning paradigm. In this method a set of training examples is broken down into smaller and smaller subsets while at the same time an associated decision tree get incrementally developed. At the end of the learning process, a decision tree covering the training set is returned. The decision tree can be thought of as a set sentences (in Disjunctive Normal Form) written propositional logic. Some characteristics of problems that are well suited to Decision Tree Learning are: Attribute-value paired elements Discrete target function Disjunctive descriptions (of target function) Works well with missing or erroneous training data

Decision Tree Learning (Outlook = Sunny  Humidity = Normal)  (Outlook = Overcast)  (Outlook = Rain  Wind = Weak)   [See: Tom M. Mitchell, Machine Learning, McGraw-Hill, 1997]

Decision Tree Learning Day Outlook Temperature Humidity Wind PlayTennis D1 Sunny Hot High Weak No D2 Strong D3 Overcast Yes D4 Rain Mild D5 Cool Normal D6 D7 D8 D9 D10 D11 D12 D13 D14 [See: Tom M. Mitchell, Machine Learning, McGraw-Hill, 1997]

Decision Tree Learning Building a Decision Tree First test all attributes and select the on that would function as the best root; Break-up the training set into subsets based on the branches of the root node; Test the remaining attributes to see which ones fit best underneath the branches of the root node; Continue this process for all other branches until all examples of a subset are of one type there are no examples left (return majority classification of the parent) there are no more attributes left (default value should be majority classification)

Decision Tree Learning Determining which attribute is best (Entropy & Gain) Entropy (E) is the minimum number of bits needed in order to classify an arbitrary example as yes or no E(S) = ci=1 –pi log2 pi , Where S is a set of training examples, c is the number of classes, and pi is the proportion of the training set that is of class i For our entropy equation 0 log2 0 = 0 The information gain G(S,A) where A is an attribute G(S,A)  E(S) - v in Values(A) (|Sv| / |S|) * E(Sv)

Decision Tree Learning Let’s Try an Example! Let E([X+,Y-]) represent that there are X positive training elements and Y negative elements. Therefore the Entropy for the training data, E(S), can be represented as E([9+,5-]) because of the 14 training examples 9 of them are yes and 5 of them are no.

Decision Tree Learning: A Simple Example Let’s start off by calculating the Entropy of the Training Set. E(S) = E([9+,5-]) = (-9/14 log2 9/14) + (-5/14 log2 5/14) = 0.94

Decision Tree Learning: A Simple Example Next we will need to calculate the information gain G(S,A) for each attribute A where A is taken from the set {Outlook, Temperature, Humidity, Wind}.

Decision Tree Learning: A Simple Example The information gain for Outlook is: G(S,Outlook) = E(S) – [5/14 * E(Outlook=sunny) + 4/14 * E(Outlook = overcast) + 5/14 * E(Outlook=rain)] G(S,Outlook) = E([9+,5-]) – [5/14*E(2+,3-) + 4/14*E([4+,0-]) + 5/14*E([3+,2-])] G(S,Outlook) = 0.94 – [5/14*0.971 + 4/14*0.0 + 5/14*0.971] G(S,Outlook) = 0.246

Decision Tree Learning: A Simple Example G(S,Temperature) = 0.94 – [4/14*E(Temperature=hot) + 6/14*E(Temperature=mild) + 4/14*E(Temperature=cool)] G(S,Temperature) = 0.94 – [4/14*E([2+,2-]) + 6/14*E([4+,2-]) + 4/14*E([3+,1-])] G(S,Temperature) = 0.94 – [4/14 + 6/14*0.918 + 4/14*0.811] G(S,Temperature) = 0.029

Decision Tree Learning: A Simple Example G(S,Humidity) = 0.94 – [7/14*E(Humidity=high) + 7/14*E(Humidity=normal)] G(S,Humidity = 0.94 – [7/14*E([3+,4-]) + 7/14*E([6+,1-])] G(S,Humidity = 0.94 – [7/14*0.985 + 7/14*0.592] G(S,Humidity) = 0.1515

Decision Tree Learning: A Simple Example G(S,Wind) = 0.94 – [8/14*0.811 + 6/14*1.00] G(S,Wind) = 0.048

Decision Tree Learning: A Simple Example Outlook is our winner!

Decision Tree Learning: A Simple Example Now that we have discovered the root of our decision tree we must now recursively find the nodes that should go below Sunny, Overcast, and Rain.

Decision Tree Learning: A Simple Example G(Outlook=Rain, Humidity) = 0.971 – [2/5*E(Outlook=Rain ^ Humidity=high) + 3/5*E(Outlook=Rain ^Humidity=normal] G(Outlook=Rain, Humidity) = 0.02 G(Outlook=Rain,Wind) = 0.971- [3/5*0 + 2/5*0] G(Outlook=Rain,Wind) = 0.971

Decision Tree Learning: A Simple Example Now our decision tree looks like: