DECISION TREES Asher Moody, CS 157B. Overview  Definition  Motivation  Algorithms  ID3  Example  Entropy  Information Gain  Applications  Conclusion.

Slides:



Advertisements
Similar presentations
Data Mining Lecture 9.
Advertisements

DECISION TREES. Decision trees  One possible representation for hypotheses.
CHAPTER 9: Decision Trees
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Paper By - Manish Mehta, Rakesh Agarwal and Jorma Rissanen
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
Hunt’s Algorithm CIT365: Data Mining & Data Warehousing Bajuna Salehe
Decision Tree Approach in Data Mining
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Part I Introduction to Data Mining by Tan,
K-NEAREST NEIGHBORS AND DECISION TREE Nonparametric Supervised Learning.
Classification: Definition Given a collection of records (training set ) –Each record contains a set of attributes, one of the attributes is the class.
Decision Tree.
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Classification Techniques: Decision Tree Learning
Overview Previous techniques have consisted of real-valued feature vectors (or discrete-valued) and natural measures of distance (e.g., Euclidean). Consider.
Lecture Notes for Chapter 4 Introduction to Data Mining
Decision Tree Rong Jin. Determine Milage Per Gallon.
ID3 Algorithm Abbas Rizvi CS157 B Spring What is the ID3 algorithm? ID3 stands for Iterative Dichotomiser 3 Algorithm used to generate a decision.
Decision Tree Algorithm
About ISoft … What is Decision Tree? Alice Process … Conclusions Outline.
Ensemble Learning: An Introduction
Induction of Decision Trees
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
Lecture 5 (Classification with Decision Trees)
Classification and Prediction by Yen-Hsien Lee Department of Information Management College of Management National Sun Yat-Sen University March 4, 2003.
Classification.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
Ordinal Decision Trees Qinghua Hu Harbin Institute of Technology
Machine Learning Chapter 3. Decision Tree Learning
Learning what questions to ask. 8/29/03Decision Trees2  Job is to build a tree that represents a series of questions that the classifier will ask of.
Decision Trees & the Iterative Dichotomiser 3 (ID3) Algorithm David Ramos CS 157B, Section 1 May 4, 2006.
Comparing Univariate and Multivariate Decision Trees Olcay Taner Yıldız Ethem Alpaydın Department of Computer Engineering Bogazici University
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Decision Trees. Decision trees Decision trees are powerful and popular tools for classification and prediction. The attractiveness of decision trees is.
For Wednesday No reading Homework: –Chapter 18, exercise 6.
CS690L Data Mining: Classification
For Monday No new reading Homework: –Chapter 18, exercises 3 and 4.
What is Data Mining? process of finding correlations or patterns among dozens of fields in large relational databases process of finding correlations or.
DATA MINING By Cecilia Parng CS 157B.
Today’s Topics Learning Decision Trees (Chapter 18) –We’ll use d-trees to introduce/motivate many general issues in ML (eg, overfitting reduction) “Forests”
ID3 Algorithm Michael Crawford.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
CS112 Scientific Computation Department of Computer Science Wellesley College AI designs a forest Decision trees.
DECISION TREE Ge Song. Introduction ■ Decision Tree: is a supervised learning algorithm used for classification or regression. ■ Decision Tree Graph:
An Introduction Student Name: Riaz Ahmad Program: MSIT( ) Subject: Data warehouse & Data Mining.
Decision Tree Learning Presented by Ping Zhang Nov. 26th, 2007.
Decision Tree Algorithms Rule Based Suitable for automatic generation.
Big Data Analysis and Mining Qinpei Zhao 赵钦佩 2015 Fall Decision Tree.
1 Classification: predicts categorical class labels (discrete or nominal) classifies data (constructs a model) based on the training set and the values.
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 4-Inducción de árboles de decisión (1/2) Eduardo Poggi.
1 By: Ashmi Banerjee (125186) Suman Datta ( ) CSE- 3rd year.
BY International School of Engineering {We Are Applied Engineering} Disclaimer: Some of the Images and content have been taken from multiple online sources.
DATA MINING TECHNIQUES (DECISION TREES ) Presented by: Shweta Ghate MIT College OF Engineering.
Presentation prepared by Yehonatan Cohen and Danny Hendler Some of the slides based on the online book “Social media mining” Danny Hendler Advanced Topics.
Ning Jin, Wei Wang ICDE 2011 LTS: Discriminative Subgraph Mining by Learning from Search History.
CSE573 Autumn /09/98 Machine Learning Administrative –Last topic: Decision Tree Learning Reading: 5.1, 5.4 Last time –finished NLP sample system’s.
DECISION TREES An internal node represents a test on an attribute.
k-Nearest neighbors and decision tree
Decision Trees.
Decision Trees (suggested time: 30 min)
Classification and Prediction
Machine Learning Week 1.
Machine Learning Chapter 3. Decision Tree Learning
Data Mining – Chapter 3 Classification
Machine Learning Chapter 3. Decision Tree Learning
©Jiawei Han and Micheline Kamber
CS639: Data Management for Data Science
Presentation transcript:

DECISION TREES Asher Moody, CS 157B

Overview  Definition  Motivation  Algorithms  ID3  Example  Entropy  Information Gain  Applications  Conclusion

Decision Tree  Decision trees are a fundamental technique used in data mining.  Decision trees are used for classification, clustering, feature selection, and prediction.

Motivation  Decision trees help accurate classify data  Decision trees help understand the predictive nature of the data by recognizing patterns  Decision trees depict the relationships between input data and target outputs

Algorithms  Decision trees algorithms are greedy so once test has been selected to partition the data other options will not be explored  Popular Algorithms  Computer Science: ID3, C4.5, and C5.0  Statistics: Classification and Regression Trees (CART)

ID3 Algorithm  Given: Examples(S); Target attribute (C); Attributes (R)  Initialize Root  Function ID3 (S,C,R)  Create a Root node for the tree  IF S = empty, return a single node with value Failure;  IF S = C, return a single node C;  IF R = empty, return a single node with most frequent target attribute (C);  ELSE  BEGIN… (next slide)

ID3 (cont)  BEGIN  Let D be the attribute with largest Gain Radio (D, S) among attributes in R;  Let {d j | j = 1, 2, …, n} be the values of attribute D;  Let {S j | j = 1, 2, …, n} be the subsets of S consisting respectively of records with value d j for attribute D;  Return a tree with root labeled D arcs d 1, d 2, …, d n going respectively to the trees;  For each branch in the tree  IF S = empty, add a new branch with most frequent C;  ELSE  ID3 (S1, C, R – {D}), ID3 (S2, C, R – {D}), …, IDC(Sn, C, R – {D})  END ID3  Return Root

Example 1

Example 2

Entropy  Entropy gives us a measure of how uncertain we are about the data  Maximum: The measure should be maximal if all the outcomes are equally likely (uncertainty is highest when all possible events are equiprobable). where Pi is the proportion of instances in the dataset that take the ith value of the target attribute

Information Gain  Gain calculates the reduction in entropy (gain in information) that would result from splitting the data at a particular attribute A. where v is a value of A, |Sv| is the subset of instances of S where A takes the value v, and |S| is the number of instances

Applications  Business: to track purchasing patterns  Medical: identify potential risks associated with diseases  Banks: identify potential credit risks  Governments: to determine features of potential terrorists  Seismology: to predict earthquakes

Conclusion  Search through attributes to find the proportions  Calculate the entropy for each possible data input for a particular attribute  Calculate the gain for each attribute  Make the attribute with the highest gain the root node  Continue the process until decision tree is complete

References  Berry, M. W. (2006). Lecture Notes in Data Mining. World Scientific    decision_trees