Presentation is loading. Please wait.

Presentation is loading. Please wait.

Decision Tree Rong Jin. Determine Milage Per Gallon.

Similar presentations


Presentation on theme: "Decision Tree Rong Jin. Determine Milage Per Gallon."— Presentation transcript:

1 Decision Tree Rong Jin

2 Determine Milage Per Gallon

3 A Decision Tree for Determining MPG From slides of Andrew Moore mpgcylindersdisplacementhorsepowerweightaccelerationmodelyearmaker 4lowlowlowhigh75to78asia good

4 Decision Tree Learning  Extremely popular method Credit risk assessment Medical diagnosis Market analysis  Good at dealing with symbolic feature  Easy to comprehend Compared to logistic regression model and support vector machine

5 Representational Power  Q: Can trees represent arbitrary Boolean expressions?  Q: How many Boolean functions are there over N binary attributes?

6 How to Generate Trees from Training Data

7 A Simple Idea  Enumerate all possible trees  Check how well each tree matches with the training data  Pick the one work best Too many trees Problems ? How to determine the quality of decision trees?

8 Solution: A Greedy Approach  Choose the most informative feature  Split data set  Recursive until each data item is classified correctly

9 How to Determine the Best Feature?  Which feature is more informative to MPG?  What metric should be used? From Andrew Moore’s slides Mutual Information !

10 Mutual Information for Selecting Best Features From Andrew Moore’s slides

11 Another Example: Playing Tennis

12 Example: Playing Tennis Humidity HighNorm (9+, 5-) (3+, 4-) (6+, 1-) Wind WeakStrong (9+, 5-) (6+, 2-) (3+, 3-)

13 Predication for Nodes From Andrew Moore’s slides What is the predication for each node?

14 Predication for Nodes

15 Recursively Growing Trees Original Dataset Partition it according to the value of the attribute we split on cylinders = 4 cylinders = 5 cylinders = 6 cylinders = 8 From Andrew Moore slides

16 Recursively Growing Trees cylinders = 4 cylinders = 5 cylinders = 6 cylinders = 8 Build tree from These records.. Build tree from These records.. Build tree from These records.. Build tree from These records.. From Andrew Moore slides

17 A Two Level Tree Recursively growing trees

18 When should We Stop Growing Trees? Should we split this node ?

19 Base Cases  Base Case One: If all records in current data subset have the same output then don’t recurse  Base Case Two: If all records have exactly the same set of input attributes then don’t recurse

20 Base Cases: An idea  Base Case One: If all records in current data subset have the same output then don’t recurse  Base Case Two: If all records have exactly the same set of input attributes then don’t recurse Proposed Base Case 3: If all attributes have zero information gain then don’t recurse Is this a good idea?

21 Old Topic: Overfitting

22 What should We do ? Pruning

23 Pruning Decision Tree  Stop growing trees in time  Build the full decision tree as before.  But when you can grow it no more, start to prune: Reduced error pruning Rule post-pruning

24 Reduced Error Pruning  Split data into training and validation set  Build a full decision tree over the training set  Keep removing node that maximally increases validation set accuracy

25 Original Decision Tree

26 Pruned Decision Tree

27 Reduced Error Pruning

28 Rule Post-Pruning  Convert tree into rules  Prune rules by removing the preconditions  Sort final rules by their estimated accuracy Most widely used method (e.g., C4.5) Other methods: statistical significance test (chi- square)

29 Real Value Inputs  What should we do to deal with real value inputs?

30 Information Gain  x: a real value input  t: split value  Find the split value t such that the mutual information I(x, y: t) between x and the class label y is maximized.

31 Conclusions  Decision trees are the single most popular data mining tool Easy to understand Easy to implement Easy to use Computationally cheap  It’s possible to get in trouble with overfitting  They do classification: predict a categorical output from categorical and/or real inputs

32 Software  Most widely used decision tree C4.5 (or C5.0) http://www2.cs.uregina.ca/~hamilton/courses/83 1/notes/ml/dtrees/c4.5/tutorial.html Source code, tutorial

33 The End


Download ppt "Decision Tree Rong Jin. Determine Milage Per Gallon."

Similar presentations


Ads by Google