Presentation is loading. Please wait.

Presentation is loading. Please wait.

Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.

Similar presentations


Presentation on theme: "Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn."— Presentation transcript:

1 Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn EmilyRedAverageHeavyNoSunburn PeteBrownTallHeavyNoNone JohnBrownAverageHeavyNoNone KatieBlondeShortLightYesNone

2 Decision Tree

3 DT Inducer

4 DayOutlookTempHumidityWindPlay? D1SunnyHotHighWeakNo D2SunnyHotHighStrongNo D3OvercastHotHighWeakYes D4RainMildHighWeakYes D5RainCoolNormalWeakYes D6RainCoolNormalStrongNo D7OvercastCoolNormalStrongYes D8SunnyMildHighWeakNo D9SunnyCoolNormalWeakYes D10RainMildNormalWeakYes D11SunnyMildNormalStrongYes D12OvercastMildHighStrongYes D13OvercastHotNormalWeakNo D14RainMildHighStrongNo What’s the DT?

5 Root  Outlook!, humidity, wind, temp

6 Version Space VS Decision Tree  ID3 searches a complete hypothesis space (any finite-valued discrete function). It searches incompletely using hill climbing with the heuristic: Preferring shorter trees with high information gain attributes closer to the root (Inductive bias)  Version spaces search an incomplete hypothesis space completely. Inductive bias arises from the bias in the hypothesis representation

7 Issues  How deep?  Continuous attributes?  Missing attribute values?  Attributes with different costs?

8 Overfitting  Tree grows deep enough to perform well on training data. But There may be noise in the data Not enough examples  Over-fitting: h overfits the data if h’ does worse on training examples but does better over all instances

9

10 Handling Overfitting  Stop growing tree earlier  Post pruning (works better in practice)

11 Methods  Construct tree then use a validation set – a separate set of examples, distinct from training set to evaluate the utility of post-pruning nodes  Construct Tree with all available data then use statistical tests to determine whether expanding or pruning a node is likely to produce an improvement only on the training example or the entire instance distribution

12 Training and validation  2/3 used for training  1/3 used for validation (Should be a large enough sample) Validation set is a safety check Validation set is unlikely to contain the same random errors and coincidental regularities as the training set.

13

14 Reduced error pruning  Pruning a node: Remove subtree rooted at that node – make it a leaf node, and give it the most common classification of training examples at that node Consider each node for pruning Remove node only if the pruned tree performs no worse than original Iterate and stop when further pruning decreases decision accuracy on validation set

15 From Trees to Rules  Traverse DT from root to each leaf Each such path defines a rule  Example  If ?x hair color is blonde ?x uses lotion  Then Nothing happens

16 Rules from DT  If ?x hair color is blonde ?x uses no lotion  Then ?x turns red  If ?x hair color is red  Then ?x turns red  If ?x hair color is blonde ?x uses lotion  Then Nothing happens  If ?x hair color is dark  Then Nothing happens

17 Rule Pruning  Eliminate unnecessary antecedents. Consider:  If ?x hair color is blonde  ?x uses lotion  Then nothing happens  Suppose we eliminate the first antecedent (blonde) The rule triggers for each person who uses lotion If ?x uses lotion Then nothing happens Data shows that nothing happens to anyone using lotion!  Might as well drop the first antecedent since it makes no difference!

18 Contingency tables  Formalizing the intuition. For those who used lotion: No changeSunburned Blonde20 Not Blonde10  For those who used lotion, it does not matter if they are blonde or not blonde, they do not get sunburned.

19 Contingency tables  Formalize the intuition. For those who are blonde: No changeSunburned lotion2 (Dane, Katie)0 No lotion02 (Sarah, Annie)  For those who are blonde, it does matter whether or not they use lotion. Two of those who use lotion get sunburned and two do not.

20 Contingency tables  If ?x is blonde and does not use lotion then ?x turns red  Eliminate ?x is blonde  if ?x does not use lotion then ?x turns red  For those who do not use lotion: No changeSunburned Blonde02 Not Blonde21  Looks like ?x is blonde is important

21 Contingency tables  If ?x is blonde and does not use lotion then ?x turns red  Eliminate ?x does not use lotion  if ?x is blonde then ?x turns red  For those who are blonde: No changeSunburned No lotion02 Lotion20  Looks like ?x does not use lotion is important

22 Contingency tables  If ?x is redhead Then ?x turns red  Eliminate ?x is redhead  Rule always fires!  Look at Everyone: No changeSunburned Redhead01 Not redhead52  Evidently red hair is important

23 Contingency tables  If ?x is dark haired Then Nothing happens  Eliminate ?x is dark haired  Rule always fires!  Look at Everyone: No changeSunburned Dark haired Not Dark haired  Is being dark haired important?

24 Contingency tables  If ?x is dark haired Then Nothing happens  Eliminate ?x is dark haired  Rule always fires!  Look at Everyone: No changeSunburned Dark haired30 Not Dark haired23  Is being dark haired important?

25 Eliminate Unnecessary rules  If ?x is blonde; ?x uses no lotion Then ?x is sunburned  If ?x uses lotion Then ?x Nothing happens  If ?x is redhead Then ?x is sunburned  If ?x is dark haired Then Nothing happens  Can we come up with a default rule that eliminates the need for some of the above rules?

26 Eliminate Unnecessary rules  If ?x uses lotion Then ?x Nothing happens  If ?x is dark haired Then Nothing happens  If no other rule applies  Then Nothing happens

27 Default rules  If ?x is blonde; ?x uses no lotion Then ?x is sunburned  If ?x is redhead Then ?x is sunburned  If no other rule applies  Then ?x gets sunburned

28 Eliminate rules using defaults  Heuristic 1: Choose the default rule that eliminates/replaces as many other rules as possible Both default rules eliminate 2 other rules  cannot use this heuristic  Heuristic 2 (to choose among rules identified by heuristic 1): Choose the default rule that covers the most common consequent 5 not sunburned, 3 sunburned Choose: If no other rule applies Then nothing happens

29 General Procedure: Fisher’s exact test cab be used to do this check

30 Why rules?  Distinguish between different contexts: Each rule can be considered separately for antecedent pruning. Contrast this with DT. You can remove a node or not? But there may be many contexts through that node. That is, there may be many rules that go through that node and removing the node means you remove ALL rules that go through that node If we consider each rule separately, we can consider all contexts of a DT node separately.

31 Why Rules?  In DTs nodes near root are more “important” than nodes near leaves.  Rules avoid the distinction between attribute tests near the root and those later on. We can avoid thinking about re-organizing the tree if we have to remove the root node!

32 Why rules?  Rules may be more readable

33 Continuous valued attributes Temp404860728090 Play?No Yes No Sort examples according to continuous values Identify adjacent examples that differ in target values (play?) Pick one of (48 + 60)/2 and (90 + 80)/2 by evaluating disorder of the new feature “tempGT54” and “tempGT85” Can also come up with multiple intervals. Use both above?

34 Feature selection  Search through space of feature subsets for a subset that maximizes performance


Download ppt "Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn."

Similar presentations


Ads by Google