Zhangxi Lin ISQS 7342-001 Texas Tech University Note: Most slides are from Decision Tree Modeling by SAS Lecture Notes 6 Ensembles of Trees.

Slides:



Advertisements
Similar presentations
When Efficient Model Averaging Out-Perform Bagging and Boosting Ian Davidson, SUNY Albany Wei Fan, IBM T.J.Watson.
Advertisements

Random Forest Predrag Radenković 3237/10
Ensemble Learning Reading: R. Schapire, A brief introduction to boosting.
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
Longin Jan Latecki Temple University
Paper presentation for CSI5388 PENGCHENG XI Mar. 23, 2005
MCS Multiple Classifier Systems, Cagliari 9-11 June Giorgio Valentini Random aggregated and bagged ensembles.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Sparse vs. Ensemble Approaches to Supervised Learning
Ensemble Learning what is an ensemble? why use an ensemble?
Ensemble Learning: An Introduction
Three kinds of learning
Bagging LING 572 Fei Xia 1/24/06. Ensemble methods So far, we have covered several learning methods: FSA, HMM, DT, DL, TBL. Question: how to improve results?
Machine Learning: Ensemble Methods
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Ensemble Learning (2), Tree and Forest
For Better Accuracy Eick: Ensemble Learning
Machine Learning CS 165B Spring 2012
Issues with Data Mining
Lecture Notes 4 Pruning Zhangxi Lin ISQS
ENSEMBLE LEARNING David Kauchak CS451 – Fall 2013.
Boosting Neural Networks Published by Holger Schwenk and Yoshua Benggio Neural Computation, 12(8): , Presented by Yong Li.
CS 391L: Machine Learning: Ensembles
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Ensemble Classification Methods Rayid Ghani IR Seminar – 9/26/00.
ECE 8443 – Pattern Recognition Objectives: Bagging and Boosting Cross-Validation ML and Bayesian Model Comparison Combining Classifiers Resources: MN:
Data Mining - Volinsky Columbia University 1 Topic 10 - Ensemble Methods.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensemble Methods: Bagging and Boosting
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
CS Ensembles1 Ensembles. 2 A “Holy Grail” of Machine Learning Automated Learner Just a Data Set or just an explanation of the problem Hypothesis.
CLASSIFICATION: Ensemble Methods
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
ISQS 6347, Data & Text Mining1 Ensemble Methods. ISQS 6347, Data & Text Mining 2 Ensemble Methods Construct a set of classifiers from the training data.
Bayesian Averaging of Classifiers and the Overfitting Problem Rayid Ghani ML Lunch – 11/13/00.
Ensemble Methods in Machine Learning
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Classification Ensemble Methods 1
Data Analytics CMIS Short Course part II Day 1 Part 3: Ensembles Sam Buttrey December 2015.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Ensemble Methods Construct a set of classifiers from the training data Predict class label of previously unseen records by aggregating predictions made.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
… Algo 1 Algo 2 Algo 3 Algo N Meta-Learning Algo.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bagging and Boosting Cross-Validation ML.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
1 Machine Learning Lecture 8: Ensemble Methods Moshe Koppel Slides adapted from Raymond J. Mooney and others.
Ensemble Classifiers.
Machine Learning: Ensemble Methods
Bagging and Random Forests
Chapter 13 – Ensembles and Uplift
COMP61011 : Machine Learning Ensemble Models
Ensemble Learning Introduction to Machine Learning and Data Mining, Carla Brodley.
A “Holy Grail” of Machine Learing
INTRODUCTION TO Machine Learning
Combining Base Learners
Bayesian Averaging of Classifiers and the Overfitting Problem
Introduction to Data Mining, 2nd Edition
Lecture 18: Bagging and Boosting
Multiple Decision Trees ISQS7342
Ensemble learning.
Lecture 06: Bagging and Boosting
Model Combination.
Ensemble learning Reminder - Bagging of Trees Random Forest
Classification with CART
INTRODUCTION TO Machine Learning 3rd Edition
CS 391L: Machine Learning: Ensembles
… 1 2 n A B V W C X 1 2 … n A … V … W … C … A X feature 1 feature 2
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Presentation transcript:

Zhangxi Lin ISQS Texas Tech University Note: Most slides are from Decision Tree Modeling by SAS Lecture Notes 6 Ensembles of Trees

Chapter 5: Ensembles of Trees 5.1 Forests 5.2 Bagged Tree Models

Chapter 5: Ensembles of Trees 5.1 Forests 5.2 Bagged Tree Models

Instability Accuracy = 80% One reversal

The Challenge Decision trees are unstable models. Small changes in the training data can cause large changes in the topology of the tree. However, the overall performance of the tree remains stable (Breiman et al. 1984). The instability results from the large number of univariate splits considered and the fragmentation of the data. At each split, there are typically a number of splits on the same and different inputs that give similar performance (competitor splits). A small change in the data can easily result in a different split being chosen. This in turn produces different subsets in the child nodes; the changes in the data are even larger in the child nodes. The changes continue to cascade down the tree.

Competitor Splits Logworth Input Range minmax X1 X2

Overcome the Instability Using perturb and combine (P & C) methods to generate multiple models by manipulating the distribution of the data or altering the construction method and then averaging the results (Breiman 1998). Any unstable modeling method can be used, but trees are most often chosen because of their speed and flexibility. Some perturbation methods: –resample –subsample –add noise –adaptively reweight –randomly choose from the competitor splits

Perturb

Ensemble Model An ensemble model is the combination of multiple models. The combinations can be formed by –voting on the classifications –weighted voting where some models have more weight –averaging (weighted or unweighted) the predicted values. The attractiveness of P & C methods is their improved performance because of variance reduction. If the base models have low bias and high variance, then averaging decreases the variance. However, combining stable models can negatively affect performance. The reasons why adaptive P & C methods work go beyond simple variance reduction and are the topic of current research. Graphical explanations show that ensembles of trees have decision boundaries of much finer resolution.

Combine T1T1 T2T2 T3T3 ave(T 1, T 2, T 3 ) = Truth

Bagging case freq k=1 freq k=2 freq k=3 freq k=4 …

Bagging (bootstrap aggregation) The original P & C method (Breiman 1996). Steps: Draw B bootstrap samples. –A bootstrap sample is a random sample of size n drawn from the empirical distribution of a sample of size n. Build a tree on each bootstrap sample. Vote or average. –For classification problems, take the mean of the posterior probabilities or take the plurality vote of the predicted class. –For regression problems, take a mean of the predicted values. For example, –Breiman (1996) used 50 bootstrap replicates for classification, and 25 for regression and for averaging the posterior probabilities. –Bauer and Kohavi (1999) used 25 replicates for both voting and averaging.

Arc-x4 case freq 1 k=1 freq k=2 freq k=3 freq k=4 … m101000m m102100m m203101m ARC: Adaptive Resampling and Combining

Single, Bagged, and Boosted Tree

Chapter 5: Ensembles of Trees 5.1 Forests 5.2 Bagged Tree Models

Bagged Tree Models The current version of SAS Enterprise Miner does not support bagging or boosting. They can be implemented in a Code node using SAS programming and the ARBORETUM procedure. The program over-fits 25 trees, each to a different 50% random sample of the training data. Validation data is not used to tune the models. Training, Validation, and Test sets are scored from within PROC ARBORETUM. The model predictions are then averaged, and profit is calculated for each set.