Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, 11-12 a Machine Learning.

Slides:



Advertisements
Similar presentations
Decision Tree Learning
Advertisements

Decision Tree Learning - ID3
Decision Trees Decision tree representation ID3 learning algorithm
Prof. Sin-Min Lee Department of Computer Science
Decision Tree Algorithm (C4.5)
ICS320-Foundations of Adaptive and Learning Systems
Classification Techniques: Decision Tree Learning
Decision Tree Example MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way.
Decision Trees Instructor: Qiang Yang Hong Kong University of Science and Technology Thanks: Eibe Frank and Jiawei Han.
Machine Learning Group University College Dublin Decision Trees What is a Decision Tree? How to build a good one…
Decision Tree Learning
Part 7.3 Decision Trees Decision tree representation ID3 learning algorithm Entropy, information gain Overfitting.
Decision Tree Learning Learning Decision Trees (Mitchell 1997, Russell & Norvig 2003) –Decision tree induction is a simple but powerful learning paradigm.
Induction of Decision Trees
1 Classification with Decision Trees I Instructor: Qiang Yang Hong Kong University of Science and Technology Thanks: Eibe Frank and Jiawei.
Decision Trees an Introduction.
Machine Learning Reading: Chapter Text Classification  Is text i a finance new article? PositiveNegative.
Machine Learning Lecture 10 Decision Trees G53MLE Machine Learning Dr Guoping Qiu1.
Theses slides are based on the slides by
ID3 Algorithm Allan Neymark CS157B – Spring 2007.
Machine Learning Chapter 3. Decision Tree Learning
Learning what questions to ask. 8/29/03Decision Trees2  Job is to build a tree that represents a series of questions that the classifier will ask of.
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
Machine Learning Lecture 10 Decision Tree Learning 1.
CpSc 810: Machine Learning Decision Tree Learning.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Machine Learning Queens College Lecture 2: Decision Trees.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Longin Jan Latecki Temple University
For Wednesday No reading Homework: –Chapter 18, exercise 6.
Data Mining – Algorithms: Decision Trees - ID3 Chapter 4, Section 4.3.
For Monday No new reading Homework: –Chapter 18, exercises 3 and 4.
Chapter 20 Data Analysis and Mining. 2 n Decision Support Systems  Obtain high-level information out of detailed information stored in (DB) transaction-processing.
1 Decision Tree Learning Original slides by Raymond J. Mooney University of Texas at Austin.
Decision Trees, Part 1 Reading: Textbook, Chapter 6.
Exercises Decision Trees In decision tree learning, the information gain criterion helps us select the best attribute to split the data at every node.
Machine Learning Decision Trees. E. Keogh, UC Riverside Decision Tree Classifier Ross Quinlan Antenna Length Abdomen Length.
DECISION TREE Ge Song. Introduction ■ Decision Tree: is a supervised learning algorithm used for classification or regression. ■ Decision Tree Graph:
Decision Tree Learning
Training Examples. Entropy and Information Gain Information answers questions The more clueless I am about the answer initially, the more information.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Presentation on Decision trees Presented to: Sir Marooof Pasha.
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
Oliver Schulte Machine Learning 726 Decision Tree Classifiers.
SEEM Tutorial 1 Classification: Decision tree Siyuan Zhang,
Iterative Dichotomiser 3 By Christopher Archibald.
Prof. Pushpak Bhattacharyya, IIT Bombay1 CS 621 Artificial Intelligence Lecture 12 – 30/08/05 Prof. Pushpak Bhattacharyya Fundamentals of Information.
1 Machine Learning Decision Trees Some of these slides are courtesy of R.Mooney, UT Austin and E. Keogh, UC Riverside.
Review of Decision Tree Learning Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
CSE573 Autumn /11/98 Machine Learning Administrative –Finish this topic –The rest of the time is yours –Final exam Tuesday, Mar. 17, 2:30-4:20.
Decision Tree Learning
DECISION TREES An internal node represents a test on an attribute.
Decision Trees an introduction.
Data Mining Decision Tree Induction
Decision Tree Learning
Decision trees (concept learnig)
Decision trees (concept learnig)
Classification Algorithms
Decision Tree Learning
Teori Keputusan (Decision Theory)
Decision Tree Saed Sayad 9/21/2018.
ID3 Algorithm.
Jianping Fan Dept of Computer Science UNC-Charlotte
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning: Lecture 3
Decision Trees Decision tree representation ID3 learning algorithm
Play Tennis ????? Day Outlook Temperature Humidity Wind PlayTennis
Machine Learning Chapter 3. Decision Tree Learning
Decision Trees Decision tree representation ID3 learning algorithm
Data Mining CSCI 307, Spring 2019 Lecture 15
Presentation transcript:

Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning

 Each branch corresponds to attribute value  Each internal node has a splitting predicate  Each leaf node assigns a classification Machine Learning

Entropy (disorder, impurity) of a set of examples, S, relative to a binary classification is: where p 1 is the fraction of positive examples in S and p 0 is the fraction of negatives. Machine Learning

If all examples are in one category, entropy is zero (we define 0  log(0)=0) If examples are equally mixed (p 1 =p 0 =0.5), entropy is a maximum of 1. Entropy can be viewed as the number of bits required on average to encode the class of an example in S where data compression (e.g. Huffman coding) is used to give shorter codes to more likely cases. For multi-class problems with c categories, entropy generalizes to: Machine Learning

Gain (S, A) = expected reduction in entropy due to sorting on A Values (A) is the set of all possible values for attribute A, Sv is the subset of S which attribute A has value v Gain(S,A) is the expected reduction in entropy caused by knowing the values of attribute A. Machine Learning

Humidity High Normal 3+,4- 6+,1- E =.985 E =.592 Machine Learning

Humidity Wind High NormalWeakStrong 3+,4- 6+, ,3- E =.985 E =.592 E =.811 E =1.0 Machine Learning

Humidity Wind High NormalWeakStrong 3+,4- 6+, ,3- E =.985 E =.592 E =.811 E =1.0 Gain(S, Humidity ) = / / = Gain(S, Wind ) = / / = Machine Learning

Outlook OvercastRain 3,7,12,13 4,5,6,10,14 3+,2- Sunny 1,2,8,9,11 4+,0-2+, Gain(S, Outlook ) = Machine Learning

Outlook Gain(S, Wind ) =0.048 Gain(S, Humidity ) =0.151 Gain(S, Temperature ) =0.029 Gain(S, Outlook ) =0.246 Machine Learning

Outlook OvercastRain 3,7,12,13 4,5,6,10,14 3+,2- Sunny 1,2,8,9,11 4+,0-2+,3- Yes?? Machine Learning

Outlook Overcast Rain 3,7,12,13 4,5,6,10,14 3+,2- Sunny 1,2,8,9,11 4+,0-2+,3- Yes??.97-(3/5) 0-(2/5) 0 = (2/5) 1 = (2/5) 1 - (3/5).92=.02 Machine Learning

Outlook Overcast Rain 3,7,12,13 4,5,6,10,14 3+,2- Sunny 1,2,8,9,11 4+,0-2+,3- YesHumidity? NormalHigh NoYes Machine Learning

Outlook Overcast Rain 3,7,12,13 4,5,6,10,14 3+,2- Sunny 1,2,8,9,11 4+,0-2+,3- YesHumidityWind NormalHigh No Yes WeakStrong No Yes Machine Learning

Person Hair Length WeightAge Class Homer0”25036 M Vardhan10”15034 F Kumar2”9010 M Lisa6”788 F Maggie4”201 F Abe1”17070 M Selma8”16041 F Sai10”18038 M Krusty6”20045 M Machine Learning

Hair Length <= 5? yes no Entropy(4F,5M) = -(4/9)log 2 (4/9) - (5/9)log 2 (5/9) = Entropy(1F,3M) = -(1/4)log 2 (1/4) - (3/4)log 2 (3/4) = Entropy(3F,2M) = -(3/5)log 2 (3/5) - (2/5)log 2 (2/5) = Gain(Hair Length <= 5) = – (4/9 * /9 * ) = Let us try splitting on Hair length Machine Learning

Weight <= 160? yes no Entropy(4F,5M) = -(4/9)log 2 (4/9) - (5/9)log 2 (5/9) = Entropy(4F,1M) = -(4/5)log 2 (4/5) - (1/5)log 2 (1/5) = Entropy(0F,4M) = -(0/4)log 2 (0/4) - (4/4)log 2 (4/4) = 0 Let us try splitting on Weight Machine Learning

age <= 40? yes no Entropy(4F,5M) = -(4/9)log 2 (4/9) - (5/9)log 2 (5/9) = Entropy(3F,3M) = -(3/6)log 2 (3/6) - (3/6)log 2 (3/6) = 1 Entropy(1F,2M) = -(1/3)log 2 (1/3) - (2/3)log 2 (2/3) = Let us try splitting on Age Machine Learning

Weight <= 160? yesno Hair Length <= 2? yes no Male Female Machine Learning

Weight <= 160? yesno Hair Length <= 2? yes no Male Female Rules to Classify Males/Females If Weight greater than 160, classify as Male Elseif Hair Length less than or equal to 2, classify as Male Else classify as Female Rules to Classify Males/Females If Weight greater than 160, classify as Male Elseif Hair Length less than or equal to 2, classify as Male Else classify as Female Machine Learning

Thank you!