Ensemble Methods.  “No free lunch theorem” Wolpert and Macready 1995.

Slides:



Advertisements
Similar presentations
On-line learning and Boosting
Advertisements

Data Mining Classification: Alternative Techniques
Longin Jan Latecki Temple University
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Ensemble Learning what is an ensemble? why use an ensemble?
2D1431 Machine Learning Boosting.
A Brief Introduction to Adaboost
Ensemble Learning: An Introduction
Adaboost and its application
Introduction to Boosting Aristotelis Tsirigos SCLT seminar - NYU Computer Science.
Examples of Ensemble Methods
Machine Learning: Ensemble Methods
Sparse vs. Ensemble Approaches to Supervised Learning
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Boosting Main idea: train classifiers (e.g. decision trees) in a sequence. a new classifier should focus on those cases which were incorrectly classified.
For Better Accuracy Eick: Ensemble Learning
Ensembles of Classifiers Evgueni Smirnov
Machine Learning CS 165B Spring 2012
CSSE463: Image Recognition Day 27 This week This week Last night: k-means lab due. Last night: k-means lab due. Today: Classification by “boosting” Today:
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Benk Erika Kelemen Zsolt
Boosting of classifiers Ata Kaban. Motivation & beginnings Suppose we have a learning algorithm that is guaranteed with high probability to be slightly.
CS Fall 2015 (© Jude Shavlik), Lecture 7, Week 3
Data Mining - Volinsky Columbia University 1 Topic 10 - Ensemble Methods.
BOOSTING David Kauchak CS451 – Fall Admin Final project.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensemble Methods: Bagging and Boosting
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
CLASSIFICATION: Ensemble Methods
ISQS 6347, Data & Text Mining1 Ensemble Methods. ISQS 6347, Data & Text Mining 2 Ensemble Methods Construct a set of classifiers from the training data.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Learning with AdaBoost
E NSEMBLE L EARNING : A DA B OOST Jianping Fan Dept of Computer Science UNC-Charlotte.
Ensemble Learning  Which of the two options increases your chances of having a good grade on the exam? –Solving the test individually –Solving the test.
1 CHUKWUEMEKA DURUAMAKU.  Machine learning, a branch of artificial intelligence, concerns the construction and study of systems that can learn from data.
CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect.
Ensemble Methods in Machine Learning
ECE 5984: Introduction to Machine Learning Dhruv Batra Virginia Tech Topics: –Ensemble Methods: Bagging, Boosting Readings: Murphy 16.4; Hastie 16.
Classification Ensemble Methods 1
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
… Algo 1 Algo 2 Algo 3 Algo N Meta-Learning Algo.
Boosting ---one of combining models Xin Li Machine Learning Course.
AdaBoost Algorithm and its Application on Object Detection Fayin Li.
1 Machine Learning Lecture 8: Ensemble Methods Moshe Koppel Slides adapted from Raymond J. Mooney and others.
Adaboost (Adaptive boosting) Jo Yeong-Jun Schapire, Robert E., and Yoram Singer. "Improved boosting algorithms using confidence- rated predictions."
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning: Ensemble Methods
2. Skin - color filtering.
Ensemble methods with Data Streams
Session 7: Face Detection (cont.)
COMP61011 : Machine Learning Ensemble Models
Ensemble Learning Introduction to Machine Learning and Data Mining, Carla Brodley.
Cost-Sensitive Learning
Asymmetric Gradient Boosting with Application to Spam Filtering
A “Holy Grail” of Machine Learing
INTRODUCTION TO Machine Learning
Combining Base Learners
Data Mining Practical Machine Learning Tools and Techniques
Cost-Sensitive Learning
Introduction to Data Mining, 2nd Edition
Machine Learning Ensemble Learning: Voting, Boosting(Adaboost)
ADABOOST(Adaptative Boosting)
Ensemble learning.
Model Combination.
Ensemble learning Reminder - Bagging of Trees Random Forest
INTRODUCTION TO Machine Learning 3rd Edition
Presentation transcript:

Ensemble Methods

 “No free lunch theorem” Wolpert and Macready 1995

 “No free lunch theorem” Wolpert and Macready 1995  Solution search also involves searching for learners

 Different algorithms

 Different parameters

 Different algorithms  Different parameters  Different input representations/features

 Different algorithms  Different parameters  Different input representations/features  Different data

 Base learner

 Diversity over accuracy

 Model combination

 Voting  Bagging  Boosting  Cascading

 Data set = [1,2,3,4,5,6,7,8,9,10]  Samples:  Input to learner 1 = [10,2,5,10,3]  Input to learner 2 = [4,5,2,7,6,3]  Input to learner 3 = [8,8,4,9,1]

 Create complementary learners

 Train successive learners on the mistakes of predecessors

 Weak learners combine to a strong learner

 Adaboost – Adaptive Boosting

 Allows for a smaller training set

 Adaboost – Adaptive Boosting  Allows for a smaller training set  Simple classifiers

 Adaboost – Adaptive Boosting  Allows for a smaller training set  Simple classifiers  Binary

 Modify probability of drawing examples from a training set based on errors

Step 3

 Demo

 Sequence classifiers by complexity

 Use classifier j+1 if classifier j doesn’t meet a confidence threshold

 Sequence classifiers by complexity  Use classifier j+1 if classifier j doesn’t meet a confidence threshold  Train cascading classifiers on instances the previous classifier is not confident about

 Sequence classifiers by complexity  Use classifier j+1 if classifier j doesn’t meet a confidence threshold  Train cascading classifiers on instances the previous classifier is not confident about  Most examples classified quickly, harder ones passed to more expensive classifiers

 Boosting and Cascading

 Object detection/tracking  Collaborative filtering  Neural networks  Optical character recognition ++  Biometrics  Data mining

 Ensemble methods are proven effective, but why?