Ensemble Methods: Bagging.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Is Random Model Better? -On its accuracy and efficiency-
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
Longin Jan Latecki Temple University
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
Sparse vs. Ensemble Approaches to Supervised Learning
Ensemble Learning: An Introduction
Three kinds of learning
Machine Learning: Ensemble Methods
Sparse vs. Ensemble Approaches to Supervised Learning
Copyright © 2004 by Jinyan Li and Limsoon Wong Rule-Based Data Mining Methods for Classification Problems in Biomedical Domains Jinyan Li Limsoon Wong.
1 Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data Presented by: Tun-Hsiang Yang.
3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Wei FENG Geo-Resources and Environment Lab, Bordeaux INP (Bordeaux Institute of Technology), France Supervisor:
Machine Learning CS 165B Spring 2012
Issues with Data Mining
Chapter 10 Boosting May 6, Outline Adaboost Ensemble point-view of Boosting Boosting Trees Supervised Learning Methods.
ENSEMBLE LEARNING David Kauchak CS451 – Fall 2013.
CS 391L: Machine Learning: Ensembles
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Ensemble Methods: Bagging and Boosting
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
CS Ensembles1 Ensembles. 2 A “Holy Grail” of Machine Learning Automated Learner Just a Data Set or just an explanation of the problem Hypothesis.
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
Ensemble Methods in Machine Learning
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Classification Ensemble Methods 1
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
24 October 2002Data Mining & Visualization1 Data Mining and Visualization Jeremy Walton NAG Ltd, Oxford.
1 Machine Learning Lecture 8: Ensemble Methods Moshe Koppel Slides adapted from Raymond J. Mooney and others.
Copyright © 2004 by Jinyan Li and Limsoon Wong Rule-Based Data Mining Methods for Classification Problems in Biomedical Domains Jinyan Li Limsoon Wong.
Data Science Credibility: Evaluating What’s Been Learned
Ensemble Classifiers.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning: Ensemble Methods
Data Mining Practical Machine Learning Tools and Techniques
DECISION TREES An internal node represents a test on an attribute.
Bagging and Random Forests
Histograms CSE 6363 – Machine Learning Vassilis Athitsos
Chapter 18 From Data to Knowledge
Zaman Faisal Kyushu Institute of Technology Fukuoka, JAPAN
HIERARCHICAL CLASSIFICATION OF DIFFERENT CROPS USING
Eco 6380 Predictive Analytics For Economists Spring 2016
Chapter 13 – Ensembles and Uplift
Trees, bagging, boosting, and stacking
Machine Learning: Ensembles
COMP61011 : Machine Learning Ensemble Models
Ensemble Learning Introduction to Machine Learning and Data Mining, Carla Brodley.
Ungraded quiz Unit 6.
ID3 Algorithm.
Data Mining Practical Machine Learning Tools and Techniques
Machine Learning Techniques for Data Mining
REMOTE SENSING Multispectral Image Classification
Introduction to Data Mining, 2nd Edition
Machine Learning Ensemble Learning: Voting, Boosting(Adaboost)
Image Information Extraction
Instance Based Learning
Lecture 18: Bagging and Boosting
Ensembles.
Ensemble learning.
Lecture 06: Bagging and Boosting
Model Combination.
Model generalization Brief summary of methods
A task of induction to find patterns
A task of induction to find patterns
CS 391L: Machine Learning: Ensembles
Data Mining CSCI 307, Spring 2019 Lecture 6
Presentation transcript:

Ensemble Methods: Bagging

4 base classifiers 8 training instances Original order 1 2 3 4 5 6 7 8 Training set 1 Training set 2 Training set 3 Training set 4

Why ensembles? Sometimes a learning algorithm is unstable, i.e., a little change in the training set causes a big change in the learned classifier. Sometimes there is substantial noise in the training set. By using an ensemble of classifiers, we don’t just depend on the decision of just one classifier. Disadvantages Time consuming Over-fitting sometimes

3 base classifiers

Combining Decision Tress and the NN Algorithm Classify the instance using the NN algorithm applied on the training instances associated with the classification nodes (leaves). Outlook sunny overcast rainy Humidity Windy high normal false true yes

        Ensemble paradigm Use m different learning styles to learn from one training data set. Combine decisions of multiple classifiers using, e.g, weighted voting. Training Data Data1 Data m Data2         Learner1 Learner2 Learner m Model1 Model2 Model m Model Combiner Final Model

Learner1 = Learner2 = … = Learner m Homogenous ensembles Use a single, learning style but manipulate training data to make it learn multiple models. Data1  Data2  …  Data m Learner1 = Learner2 = … = Learner m Different methods for changing training data: Bagging: Resample training data with replacement Boosting: Weigh individual training vectors In WEKA, Classify=>Choose=>classifiers=>meta They take a learning algorithm as an argument (base classifier) and create a meta-classifier.

Bag size Original training set size: n No. of independent base classifiers: m For each base classifier, randomly drawing n’ examples from the original data, with replacement n’ usually < n If n=n’, on average it will contain 63.2% unique training examples. The rest are duplicates. Combine the m resulting models using simple majority vote. Decreases overall error by decreasing the variance in the results due to unstable learners, algorithms (like decision trees) whose output can change dramatically when the training data is slightly changed.

Bagging example: 2 classes

Class boundary by 1 decision tree

Boundary by 100 trees

Satellite Images Data Generated by NASA http://archive.ics.uci.edu/ml/datasets/Statlog+(Landsat+Satellite) Generated by NASA Own by the Australian Centre for Remote Sensing One frame of Landsat imagery consists of 4 digital images of the same scene in 4 different spectral (wavelength) bands. Two of these are in the visible region: green and red Two are in the near infra-red A pixel in the image corresponds to 80m by 80m of real land Pixel value = spectral band intensity Pixel value = 0 means darkest Pixel value = 255 means brightest

Record format Example: 92 115 120 94 84 102 106 79 84 102 102 83 101 126 133 103 92 112 118 85 84 103 104 81 102 126 134 104 88 121 128 100 84 107 113 87 3 Each line of data corresponds to a 3x3 square neighborhood of pixels Example: 92 115 120 94 84 102 106 79 84 102 102 83 101 126 133 103 92 112 118 85 84 103 104 81 102 126 134 104 88 121 128 100 84 107 113 87 3 Each line contains the pixel values in the 4 spectral bands (3x3)x4 = 36 numbers The last number indicates type of land The records are given in random order so that you cannot reconstruct the original landscape

Class labels Class no. Type of land 1 red soil 2 cotton crop 3 grey soil 4 damp grey soil 5 soil with vegetation stubble 6 mixture type 7 very damp grey soil There are no examples with class 6 in this particular dataset. The classification for each pixel was performed on the basis of an actual site visit by Ms. Karen Hall, when working for Professor John A. Richards, at the Centre for emote Sensing at the University of New South Wales, Australia.

Weka’s bagging Single classifier Use satellite image training and test data Classify test data using NaiveBayesSimple Observe the outputs Bagging Classify=>Choose=>meta=>Bagging Set bagSizePercent to 80 Try numIterations = 80 Observe error rate Try numIterations = 90

Misclassification rates CART: Classification And Regression Tree