BY International School of Engineering {We Are Applied Engineering} Disclaimer: Some of the Images and content have been taken from multiple online sources.

Slides:



Advertisements
Similar presentations
Chapter 7 Classification and Regression Trees
Advertisements

DECISION TREES. Decision trees  One possible representation for hypotheses.
CHAPTER 9: Decision Trees
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Classification Algorithms
Data Mining Techniques: Classification. Classification What is Classification? –Classifying tuples in a database –In training set E each tuple consists.
IT 433 Data Warehousing and Data Mining
Decision Tree Approach in Data Mining
Introduction Training Complexity, Pruning CART vs. ID3 vs. C4.5
1 Data Mining Classification Techniques: Decision Trees (BUSINESS INTELLIGENCE) Slides prepared by Elizabeth Anglo, DISCS ADMU.
Decision Tree.
Classification Techniques: Decision Tree Learning
Chapter 7 – Classification and Regression Trees
Chapter 7 – Classification and Regression Trees
Lecture outline Classification Decision-tree classification.
Decision Tree Rong Jin. Determine Milage Per Gallon.
Induction of Decision Trees
1 Classification with Decision Trees I Instructor: Qiang Yang Hong Kong University of Science and Technology Thanks: Eibe Frank and Jiawei.
Classification Continued
Example of a Decision Tree categorical continuous class Splitting Attributes Refund Yes No NO MarSt Single, Divorced Married TaxInc NO < 80K > 80K.
Classification II.
ICS 273A Intro Machine Learning
Classification.
Ensemble Learning (2), Tree and Forest
Microsoft Enterprise Consortium Data Mining Concepts Introduction to Directed Data Mining: Decision Trees Prepared by David Douglas, University of ArkansasHosted.
Introduction to Directed Data Mining: Decision Trees
PUBLIC: A Decision Tree Classifier that Integrates Building and Pruning RASTOGI, Rajeev and SHIM, Kyuseok Data Mining and Knowledge Discovery, 2000, 4.4.
DATA MINING : CLASSIFICATION. Classification : Definition  Classification is a supervised learning.  Uses training sets which has correct answers (class.
Fall 2004 TDIDT Learning CS478 - Machine Learning.
Machine Learning Chapter 3. Decision Tree Learning
Mohammad Ali Keyvanrad
1 Data Mining Lecture 3: Decision Trees. 2 Classification: Definition l Given a collection of records (training set ) –Each record contains a set of attributes,
Chapter 9 – Classification and Regression Trees
Chapter 4 Classification. 2 Classification: Definition Given a collection of records (training set ) –Each record contains a set of attributes, one of.
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
Decision Tree Learning Debapriyo Majumdar Data Mining – Fall 2014 Indian Statistical Institute Kolkata August 25, 2014.
Business Intelligence and Decision Modeling Week 9 Customer Profiling Decision Trees (Part 2) CHAID CRT.
CS690L Data Mining: Classification
Decision Trees. What is a decision tree? Input = assignment of values for given attributes –Discrete (often Boolean) or continuous Output = predicated.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 5-Inducción de árboles de decisión (2/2) Eduardo Poggi.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Decision Trees Example of a Decision Tree categorical continuous class Refund MarSt TaxInc YES NO YesNo Married Single, Divorced < 80K> 80K Splitting.
Decision Tree Learning
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Machine Learning: Decision Trees Homework 4 assigned courtesy: Geoffrey Hinton, Yann LeCun, Tan, Steinbach, Kumar.
Decision Trees.
Classification and Regression Trees
DECISION TREES Asher Moody, CS 157B. Overview  Definition  Motivation  Algorithms  ID3  Example  Entropy  Information Gain  Applications  Conclusion.
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Classification Tree Interaction Detection. Use of decision trees Segmentation Stratification Prediction Data reduction and variable screening Interaction.
Decision Tree Learning DA514 - Lecture Slides 2 Modified and expanded from: E. Alpaydin-ML (chapter 9) T. Mitchell-ML.
Review of Decision Tree Learning Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
Introduction to Machine Learning and Tree Based Methods
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
C4.5 - pruning decision trees
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Ch9: Decision Trees 9.1 Introduction A decision tree:
Introduction to Data Mining, 2nd Edition by
Introduction to Data Mining, 2nd Edition by
Introduction to Data Mining, 2nd Edition by
MIS2502: Data Analytics Classification using Decision Trees
Data Mining – Chapter 3 Classification
Decision Trees By Cole Daily CSCI 446.
Statistical Learning Dong Liu Dept. EEIS, USTC.
Classification with CART
MIS2502: Data Analytics Classification Using Decision Trees
STT : Intro. to Statistical Learning
Presentation transcript:

BY International School of Engineering {We Are Applied Engineering} Disclaimer: Some of the Images and content have been taken from multiple online sources and this presentation is intended only for knowledge sharing but not for any commercial business intention

OVERVIEW DEFINITION OF DECISION TREE WHY DECISION TREE? DECISION TREE TERMS EASY EXAMPLE CONSTRUCTING A DECISION TREE CALCULATION OF ENTROPY ENTROPY TERMINATION CRITERIA PRUNING TREES APPROACHES TO PRUNE TREE DECISION TREE ALGORITHMS LIMITATIONS ADVANTAGES VIDEO OF CONSTRUCTING A DECISION TREE

DEFINITION OF ‘DECISION TREE '  A decision tree is a natural and simple way of inducing following kind of rules. If (Age is x) and (income is y) and (family size is z) and (credit card spending is p) then he will accept the loan  It is powerful and perhaps most widely used modeling technique of all  Decision trees classify instances by sorting them down the tree from the root to some leaf node, which provides the classification of the instance

WHY DECISION TREE? Source: Decision Trees To Classify Response variable has only two categories Use standard classification tree Response variable has multiple categories Use c4.5 implementation To Predict Response variable is continuous Linear relationships between predictors and response Use standard Regression tree Nonlinear relationships between predictors and response Use c4.5 implementation

DECISION TREE TERMS Root Node Condition Check Leaf Node(Decision Point) Condition Check Branch

EASY EXAMPLE  Joe’s garage is considering hiring another mechanic.  The mechanic would cost them an additional $50,000 / year in salary and benefits.  If there are a lot of accidents in Iowa City this year, they anticipate making an additional $75,000 in net revenue.  If there are not a lot of accidents, they could lose $20,000 off of last year’s total net revenues.  Because of all the ice on the roads, Joe thinks that there will be a 70% chance of “a lot of accidents” and a 30% chance of “fewer accidents”.  Assume if he doesn’t expand he will have the same revenue as last year.

Joe’s Garage Hiring a mechanic Hire a new mechanic Cost = $50,000 70% of an chance increase in accidents Profit = $70,000 30% of a chance decrease in accidents Profit = -$20,000 Don’t hire a mechanic Cost = $0 Estimated value of “Hire Mechanic” = NPV =.7(70,000) +.3(- $20,000) - $50,000 = - $7,000 Therefore you should not hire the mechanic continued

CONSTRUCTING A DECISION TREE  Which attribute to choose?  Information Gain  ENTROPY  Where to stop?  Termination criteria Two Aspects

CALCULATION OF ENTROPY  Entropy is a measure of uncertainty in the data Entropy(S) = ∑ (i=1 to l) -|S i |/|S| * log 2 (|S i |/|S|)  S = set of examples  S i = subset of S with value v i under the target attribute  l = size of the range of the target attribute

ENTROPY  Let us say, I am considering an action like a coin toss. Say, I have five coins with probabilities for heads 0, 0.25, 0.5, 0.75 and 1. When I toss them which one has highest uncertainty and which one has the least? H = − log2  Information gain = Entropy of the system before split – Entropy of the system after split

ENTROPY: MEASURE OF RANDOMNESS

TERMINATION CRITERIA  All the records at the node belong to one class  A significant majority fraction of records belong to a single class  The segment contains only one or very small number of records  The improvement is not substantial enough to warrant making the split

PRUNING TREES  The decision trees can be grown deeply enough to perfectly classify the training examples which leads to overfitting when there is noise in the data  When the number of training examples is too small to produce a representative sample of the true target function.  Practically, pruning is not important for classification

APPROACHES TO PRUNE TREE  Three approaches –Stop growing the tree earlier, before it reaches the point where it perfectly classifies the training data, –Allow the tree to over fit the data, and then post-prune the tree. –Allow the tree to over fit the data, transform the tree to rules and then post-prune the rules.

TWO MOST POPULAR DECISION TREE ALGORITHMS  Cart –Binary split –Gini index –Cost complexity pruning  C5.0 –Multi split –Info gain –pessimistic pruning

LIMITATIONS  Class imbalance  When there are more records and very less number of attributes/features

ADVANTAGES  They are fast  Robust  Requires very little experimentation  You may also build some intuitions about your customer base. E.g. “Are customers with different family sizes truly different?

For Detailed Description on CONSTRUCTING A DECISION TREE with example Check out our video

Plot no 63/A, 1 st Floor, Road No 13, Film Nagar, Jubilee Hills, Hyderabad For Individuals (+91) /62 For Corporates (+91) Facebook: Slide share: International School of Engineering