CLASSIFICATION: Ensemble Methods

Slides:



Advertisements
Similar presentations
Ensemble Learning – Bagging, Boosting, and Stacking, and other topics
Advertisements

Ensemble Learning Reading: R. Schapire, A brief introduction to boosting.
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
Longin Jan Latecki Temple University
Bayesian Learning Rong Jin. Outline MAP learning vs. ML learning Minimum description length principle Bayes optimal classifier Bagging.
Sparse vs. Ensemble Approaches to Supervised Learning
Ensemble Learning what is an ensemble? why use an ensemble?
2D1431 Machine Learning Boosting.
Ensemble Learning: An Introduction
Three kinds of learning
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Machine Learning: Ensemble Methods
Sparse vs. Ensemble Approaches to Supervised Learning
Ensemble Learning (2), Tree and Forest
For Better Accuracy Eick: Ensemble Learning
Ensembles of Classifiers Evgueni Smirnov
Machine Learning CS 165B Spring 2012
Issues with Data Mining
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
ENSEMBLE LEARNING David Kauchak CS451 – Fall 2013.
Lecture 7 Ensemble Algorithms MW 4:00PM-5:15PM Dr. Jianjun Hu CSCE822 Data Mining and Warehousing University of South.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Ensemble Classification Methods Rayid Ghani IR Seminar – 9/26/00.
ECE 8443 – Pattern Recognition Objectives: Bagging and Boosting Cross-Validation ML and Bayesian Model Comparison Combining Classifiers Resources: MN:
Benk Erika Kelemen Zsolt
Data Mining - Volinsky Columbia University 1 Topic 10 - Ensemble Methods.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensemble Methods: Bagging and Boosting
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
CS Ensembles1 Ensembles. 2 A “Holy Grail” of Machine Learning Automated Learner Just a Data Set or just an explanation of the problem Hypothesis.
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
ISQS 6347, Data & Text Mining1 Ensemble Methods. ISQS 6347, Data & Text Mining 2 Ensemble Methods Construct a set of classifiers from the training data.
Bayesian Averaging of Classifiers and the Overfitting Problem Rayid Ghani ML Lunch – 11/13/00.
E NSEMBLE L EARNING : A DA B OOST Jianping Fan Dept of Computer Science UNC-Charlotte.
Ensemble Methods in Machine Learning
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Classification Ensemble Methods 1
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Ensemble Methods Construct a set of classifiers from the training data Predict class label of previously unseen records by aggregating predictions made.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
Machine Learning in Practice Lecture 24 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bagging and Boosting Cross-Validation ML.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
1 Machine Learning: Ensemble Methods. 2 Learning Ensembles Learn multiple alternative definitions of a concept using different training data or different.
1 Machine Learning Lecture 8: Ensemble Methods Moshe Koppel Slides adapted from Raymond J. Mooney and others.
Combining Bagging and Random Subspaces to Create Better Ensembles
Ensemble Classifiers.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning: Ensemble Methods
Data Mining Practical Machine Learning Tools and Techniques
Chapter 13 – Ensembles and Uplift
COMP61011 : Machine Learning Ensemble Models
Ensemble Learning Introduction to Machine Learning and Data Mining, Carla Brodley.
ECE 5424: Introduction to Machine Learning
A “Holy Grail” of Machine Learing
Bayesian Averaging of Classifiers and the Overfitting Problem
Data Mining Practical Machine Learning Tools and Techniques
Introduction to Data Mining, 2nd Edition
Ensembles.
Ensemble learning.
Model Combination.
Ensemble learning Reminder - Bagging of Trees Random Forest
Classification with CART
Ch13. Ensemble method (draft)
Presentation transcript:

CLASSIFICATION: Ensemble Methods Combines multiple models Construct multiple classifiers from training set Aggregate their predictions on testing set Meta-algorithm

CLASSIFICATION: Ensemble Methods Improves stability and accuracy Reduces variance Helps avoid overfitting Compensates for poor learning algorithms Uses more computation

ENSEMBLE METHODS: Examples Bagging (bootstrap aggregation) Bagging with MetaCost Random forests Boosting Stacked generalization Usually used on different learning algorithms Bayesian model combination

ENSEMBLE METHODS: Bagging Randomly create samples (with replacement) from a data set Create classifiers (same type) for each sample Run classifiers on testing sample Use majority voting to determine classification of testing sample

ENSEMBLE METHODS: Bagging with MetaCost Used when each model can output probability estimates Probability estimates used to obtain expected cost of each prediction Classifies training instances to minimize the expected cost Learns new classifier

ENSEMBLE METHODS: Random Forests Modification of applying bagging to tree learners Uses only random subsets of features at each split Promotes tree diversity

ENSEMBLE METHODS: Boosting Seeks models that complement one another Combines models of same type New models constructed to better handle those instances incorrectly handled by previous models – focuses on hard to classify examples Uses weighted averaging often adaptively

ENSEMBLE METHODS: Stacked Generalization Introduced by David Wolpert, 1992 Other algorithms trained from training set Stacking (“level-1”) algorithm uses predicitions from base (“level-0”) algorithms as inputs

ENSEMBLE METHODS: Stacked Generalization Employs j-fold cross validation of training set Train and test each of the level-0 algorithms using the split training data to create the level-0 models Test each model on each split to create level-1 data

ENSEMBLE METHODS: Stacked Generalization

ENSEMBLE METHODS: Stacked Generalization Can be used for both supervised and unsupervised learning Best performers in Netflix competition were forms of stacked generalization Can even create multiple levels of stacking(“level-2”, etc.)

ENSEMBLE METHODS: Stacked Generalization Best performers in Netflix competition were forms of stacked generalization Can even create multiple layers (“stacked stacking”) Works best with class probabilities (Tang and Witten, 1999)

ENSEMBLE METHODS: Bayesian Model Combination Built upon Bayes Model Averaging and Bayes Optimal Classifier Bayes Optimal Classifier Ensemble (using Bayes’ rule) of all hypotheses in hypothesis space On average, it is the ideal ensemble

ENSEMBLE METHODS: Bayesian Model Combination Bayes Model Averaging Approximates Bayes optimal classifer Samples from hypothesis space Monte Carlo sampling Tends to promote overfitting Performs worse in practice than simpler techniques (eg bagging)

ENSEMBLE METHODS: Bayesian Model Combination Bayes Model Combination Correction to Bayes Model Averaging Uses model weightings to create samples Overcomes drawback of BMA giving weight to single model Better performance than BMA or bagging