Learning: Identification Trees Larry M. Manevitz All rights reserved.

Slides:



Advertisements
Similar presentations
Introduction to Artificial Intelligence CS440/ECE448 Lecture 21
Advertisements

DECISION TREES. Decision trees  One possible representation for hypotheses.
Enhancements to basic decision tree induction, C4.5
Evolutionary Computing Systems Lab (ECSL), University of Nevada, Reno 1.
Chapter 5 Fundamental Algorithm Design Techniques.
Iterative Dichotomiser 3 (ID3) Algorithm Medha Pradhan CS 157B, Spring 2007.
Managerial Decision Modeling with Spreadsheets
Part 3 Probabilistic Decision Models
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 18-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter.
Machine Learning Decision Trees. Exercise Solutions.
Machine learning learning... a fundamental aspect of intelligent systems – not just a short-cut to kn acquisition / complex behaviour.
Information Extraction Lecture 6 – Decision Trees (Basic Machine Learning) CIS, LMU München Winter Semester Dr. Alexander Fraser, CIS.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
Machine Learning Group University College Dublin Decision Trees What is a Decision Tree? How to build a good one…
ID3 Algorithm Abbas Rizvi CS157 B Spring What is the ID3 algorithm? ID3 stands for Iterative Dichotomiser 3 Algorithm used to generate a decision.
Decision Tree Learning
Ensemble Learning: An Introduction
CD-ROM Chap 14-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition CD-ROM Chapter 14 Introduction.
Decision Trees Chapter 18 From Data to Knowledge.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
MACHINE LEARNING. 2 What is learning? A computer program learns if it improves its performance at some task through experience (T. Mitchell, 1997) A computer.
MACHINE LEARNING. What is learning? A computer program learns if it improves its performance at some task through experience (T. Mitchell, 1997) A computer.
Discrete Random Variables: The Binomial Distribution
Information Theory, Classification & Decision Trees Ling 572 Advanced Statistical Methods in NLP January 5, 2012.
Chapter 8 What’s she like?
Learning Chapter 18 and Parts of Chapter 20
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
ID3 Algorithm Allan Neymark CS157B – Spring 2007.
Decision Trees Advanced Statistical Methods in NLP Ling572 January 10, 2012.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
EIE426-AICV1 Machine Learning Filename: eie426-machine-learning-0809.ppt.
Decision Theory Decision theory problems are characterized by the following: 1.A list of alternatives. 2.A list of possible future states of nature. 3.Payoffs.
Machine Learning Queens College Lecture 2: Decision Trees.
1 Learning Chapter 18 and Parts of Chapter 20 AI systems are complex and may have many parameters. It is impractical and often impossible to encode all.
Math 3680 Lecture #2 Mean and Standard Deviation.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 27, 2012.
Huffman coding Content 1 Encoding and decoding messages Fixed-length coding Variable-length coding 2 Huffman coding.
Learning with Decision Trees Artificial Intelligence CMSC February 20, 2003.
ID3 Algorithm Michael Crawford.
Searching by Authority Artificial Intelligence CMSC February 12, 2008.
Princeton University COS 423 Theory of Algorithms Spring 2001 Kevin Wayne Average Case Analysis.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Exercises Decision Trees In decision tree learning, the information gain criterion helps us select the best attribute to split the data at every node.
Finding the Density for layers.  Check out this picture from your book. Which layer has the highest density?  Which layer has the lowest density? 
Decision Theory McGraw-Hill/Irwin Copyright © 2012 by The McGraw-Hill Companies, Inc. All rights reserved.
Copyright © by Holt, Rinehart and Winston. All rights reserved. Section 1 Mendel and His Peas Gregor Mendel: Who was he? Born 1822 in Austria Monk and.
ID3 Algorithm Amrit Gurung. Classification Library System Organise according to special characteristics Faster retrieval New items sorted easily Related.
Presentation on Decision trees Presented to: Sir Marooof Pasha.
Hit “enter” to advance each slide of your sample “Train the Brain” drill.
Ch1 Larson/Farber 1 1 Elementary Statistics Larson Farber Introduction to Statistics As you view these slides be sure to have paper, pencil, a calculator.
Learning with Decision Trees Artificial Intelligence CMSC February 18, 2003.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
HAIR TYPE BrownBlondBlackRedTOTALS Wavy Straight TOTALS HAIR COLOR What is the probability that a randomly selected child will have.
Introduction and Data Collection Basic Business Statistics 10 th Edition.
Suppose a sequence begins 2, 4, …………………. Would could the formula be?
Related Sample t-test A related sample t-test is used when participants in the two groups are not independent of participants in the other groups. – Repeated.
Medical Decision Making Learning: Decision Trees Artificial Intelligence CMSC February 10, 2005.
Inna Levitskaya.  Productiveness of successful people  The rule of “6 P”  The decision tree  The most important - identify the main  The word "no“
Homework 5 Corrections: Only need to compute Sum-Squared-Error and Average Entropy of clusters, not “cohesion” and “separation”
CIS, LMU München Winter Semester Dr. Alexander Fraser, CIS
Iterative Dichotomiser 3 (ID3) Algorithm
Descriptive Statistics
Decision Trees.
Prof. Dr. Alexander Fraser, CIS
Errors with Continuous data
Advanced Associative Structures
Learning with Identification Trees
Computer Vision Chapter 4
Learning Chapter 18 and Parts of Chapter 20
Errors with Continuous data
Presentation transcript:

Learning: Identification Trees Larry M. Manevitz All rights reserved

Given data; Find rules characterizing Problems –Multi-Dimensional –Want rules that generalize

Sample Data (Winston) NameHairHeightWeightLotionBurnt? Sarahblondeaverlightnoyes Danablondetallaveryesno Alexbrownshortaveryesno Annieblondeshortavernoyes Emilyredaverheavynoyes Petebrowntallheavyno Johnbrownaverheavyno Katieblondeshortlightyesnone

How to analyze data? Make identical match (unlikely) Find best match (other techniques) Build Decision Tree that –gives correct decisions for all data –is simplest (for generalizability)

Decision Trees blonde red brown Emily (yes) Alex Pete John (All No) Hair color Lotion? Yes Sarah Annie (all yes) Dana Katie (All No)

Another Decision Tree Height Hair Color Weight Hair Color

Want to make homogeneous sets? Idea: Minimize Disorder by Tests Can be measured by “ average entropy ” Formula from information theory

Disorder or Entropy Suppose two classes perfectly balanced –Then Entropy = 1 (highest value) Suppose all members in one class then entropy is 0. (note 0 log 0 = 0)

So on our tabular example, we compute the entropy result from each test and then choose the one with lowest average entropy. Test Entropy Hair0.5 Height0.69 Lotion0.61 Weight0.94

Summary Figure out the disorder (or entropy) for each possible test. Choose the one with smallest entropy Then continue on each subbranch