3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Wei FENG Geo-Resources and Environment Lab, Bordeaux INP (Bordeaux Institute of Technology), France Supervisor:

Slides:



Advertisements
Similar presentations
Random Forest Predrag Radenković 3237/10
Advertisements

Ensemble Learning Reading: R. Schapire, A brief introduction to boosting.
Data Mining and Machine Learning
My name is Dustin Boswell and I will be presenting: Ensemble Methods in Machine Learning by Thomas G. Dietterich Oregon State University, Corvallis, Oregon.
Molecular Biomedical Informatics 分子生醫資訊實驗室 Machine Learning and Bioinformatics 機器學習與生物資訊學 Machine Learning & Bioinformatics 1.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Other Classification Techniques 1.Nearest Neighbor Classifiers 2.Support Vector Machines.
Data Mining Classification: Alternative Techniques
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
Paper presentation for CSI5388 PENGCHENG XI Mar. 23, 2005
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
Ensemble Learning what is an ensemble? why use an ensemble?
2D1431 Machine Learning Boosting.
Ensemble Learning: An Introduction
Adaboost and its application
Three kinds of learning
Machine Learning: Ensemble Methods
Boosting Main idea: train classifiers (e.g. decision trees) in a sequence. a new classifier should focus on those cases which were incorrectly classified.
Rotation Forest: A New Classifier Ensemble Method 交通大學 電子所 蕭晴駿 Juan J. Rodríguez and Ludmila I. Kuncheva.
For Better Accuracy Eick: Ensemble Learning
Machine Learning CS 165B Spring 2012
AdaBoost Robert E. Schapire (Princeton University) Yoav Freund (University of California at San Diego) Presented by Zhi-Hua Zhou (Nanjing University)
1 Mining with Noise Knowledge: Error Awareness Data Mining Xindong Wu Department of Computer Science University of Vermont, USA; Hong Kong Polytechnic.
Zhangxi Lin ISQS Texas Tech University Note: Most slides are from Decision Tree Modeling by SAS Lecture Notes 6 Ensembles of Trees.
Boosting Neural Networks Published by Holger Schwenk and Yoshua Benggio Neural Computation, 12(8): , Presented by Yong Li.
CS 391L: Machine Learning: Ensembles
1 CS 391L: Machine Learning: Experimental Evaluation Raymond J. Mooney University of Texas at Austin.
Ensemble Based Systems in Decision Making Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: IEEE CIRCUITS AND SYSTEMS MAGAZINE 2006, Q3 Robi.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Ensemble Methods: Bagging and Boosting
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
CLASSIFICATION: Ensemble Methods
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
ISQS 6347, Data & Text Mining1 Ensemble Methods. ISQS 6347, Data & Text Mining 2 Ensemble Methods Construct a set of classifiers from the training data.
Remote Sensing Classification Accuracy
Co-operative Training in Classifier Ensembles Rozita Dara PAMI Lab University of Waterloo.
Ensemble Methods in Machine Learning
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Classification Ensemble Methods 1
COMP24111: Machine Learning Ensemble Models Gavin Brown
Using Classification Trees to Decide News Popularity
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Ensemble Methods Construct a set of classifiers from the training data Predict class label of previously unseen records by aggregating predictions made.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Genetic Algorithms (in 1 Slide) l GA: based on an analogy to biological evolution l Each.
… Algo 1 Algo 2 Algo 3 Algo N Meta-Learning Algo.
GLC 2000 Workshop March 2003 Land cover map of southern hemisphere Africa using SPOT-4 VEGETATION data Ana Cabral 1, Maria J.P. de Vasconcelos 1,2,
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
U.S. Department of the Interior U.S. Geological Survey Chandra Giri Ying Zhong 7/15/2015 Cropland Extent Mapping in South America.
1 Machine Learning: Ensemble Methods. 2 Learning Ensembles Learn multiple alternative definitions of a concept using different training data or different.
1 Machine Learning Lecture 8: Ensemble Methods Moshe Koppel Slides adapted from Raymond J. Mooney and others.
Ensemble Classifiers.
Machine Learning: Ensemble Methods
Reading: R. Schapire, A brief introduction to boosting
Zaman Faisal Kyushu Institute of Technology Fukuoka, JAPAN
HIERARCHICAL CLASSIFICATION OF DIFFERENT CROPS USING
Trees, bagging, boosting, and stacking
COMP61011 : Machine Learning Ensemble Models
Ensemble Learning Introduction to Machine Learning and Data Mining, Carla Brodley.
Cost-Sensitive Learning
Asymmetric Gradient Boosting with Application to Spam Filtering
Combining Base Learners
Data Mining Practical Machine Learning Tools and Techniques
Cost-Sensitive Learning
Introduction to Data Mining, 2nd Edition
Ensembles.
Ensemble learning.
Model Combination.
Ensemble Methods: Bagging.
Presentation transcript:

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Wei FENG Geo-Resources and Environment Lab, Bordeaux INP (Bordeaux Institute of Technology), France Supervisor: Samia BOUKIR CLASSIFICATION OF SATELLITE IMAGES USING MARGIN-BASED ENSEMBLE METHODS. APPLICATION TO LAND COVER MAPPING OF NATURAL SITES

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Outline Context and objectives Ensemble learning and margin theory Mislabeled training data identification and filtering based on ensemble margin Conclusions 2

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Objective = Multiple classifier framework, based on ensemble margin, to effectively and efficiently map remote sensing data Major challenges in ensemble learning  Training data class imbalance  Training data redundancy  Training data class mislabeling 3 Context and Objectives ICIP 2013 September 2013, Melbourne, Australia 3 Forest imageLand cover map

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 I. Ensemble learning and margin theory

3 ème Journée Doctorale G&E, Bordeaux, Mars Condorcet theorem (1785): even if the members of a group have just 50% of chance to individually take the right decision, a majority voting of the same group has nearly 100% of chance to take the right decision! First use of ensemble concept in machine learning: Hansen & Salamon (IEEE PAMI 1990) Classification by Random Forests (decision tree- based ensemble): Breiman (Machine Learning 2001) Marquis de Condorcet French mathematician and philosopher ( ) Introduction to ensemble learning

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Typical ensemble method  phases   Production of multiple homogeneous classifiers, and  Their combination 6 Introduction to ensemble learning Typical ensemble creation method = bagging  bootstrap sampling (with replacement) over training data to produce diverse classifiers  components of ensemble Typical multiple classifiers combination = majority voting  ensemble decision = class with most votes 6

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Margin of ensemble methods Difference between votes to different classes Classification confidence One popular ensemble margin = difference between the fraction of classifiers voting correctly and incorrectly 7

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 II. Mislabeled training data identification and filtering based on ensemble margin

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Mislabeling problem in machine learning I am confused!! Mislabeling error 9 I am a cow!

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Ensemble-based class noise identification Typical class noise filtering approach: Majority vote filter (Brodley et al ) Principle: 10 If more than half (>50%) of all the base classifiers of the ensemble classify an instance incorrectly, then this instance is tagged as mislabeled. Weakness: It identifies all the clean training instances that have been wrongly classified by the ensemble classifier as mislabeled  false positives

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Margin-based mislabeled instance elimination algorithm Approach: noise ranking-based Noise evaluation function: N(x i ) = |margin (x i )|  (x i,y i )  S / C(x i ) ≠ y i 1.Construct an ensemble classifier C with all training data (x i,y i )  S 2.Compute the margin of each training instance x i 3.Order all the training instances x i, that have been misclassified, according to their noise evaluation values N(x i ), in descending order 4.Eliminate the first M most likely mislabeled instances x i to form a new cleaner training set 5.Evaluate the cleaned training set by classification performance, on a validation set V 6.Select the best filtered training set Algorithm 11

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Margin-based mislabeled instance correction algorithm Approach: noise ranking-based Noise evaluation function: N(x i ) = |margin (x i )|  (x i,y i )  S / C(x i ) ≠ y i 1.Construct an ensemble classifier C with all training data (x i,y i )  S 2.Compute the margin of each training instance x i 3.Order all the training instances x i, that have been misclassified, according to their noise evaluation values N(x i ), in descending order 4.Correct the labels of first M most likely mislabeled instances x i using the predicted labels to form a new cleaner training set. 5.Evaluate the repaired training set by classification performance, on a validation set V 6.Select the best corrected training set. Algorithm 12

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Margin-based mislabeled instance identification results Data sets  Three remote sensing datasets for land cover mapping of sites of different types Data setTrainingValidationTestVariablesClasses Forest Urban Agriculture Table 1. Data sets

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Margin-based mislabeled instance identification results Noise filtering performance  Noise-sensitive ensemble classifier = Boosting  Two types of class noise: Random noise = 20% (training and validation sets with a percentage of randomly mislabeled instances) Actual noise = unknown (amount and type)  Noise filter strategy: adaptive filtering ADAPTIVE FILTERING eliminates or corrects an amount of ordered detected mislabeled instances = amount that led to maximum accuracy on validation set 14

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Margin-based mislabeled instance identification results: Artificial noise Table 2. Overall accuracies (%) of boosting classifier with no filter, majority vote filtered and margin-based filtered training sets on artificially corrupted data sets (noise rate=20%) 15 Increase in accuracy of up to 5%

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Margin-based mislabeled instance identification results: Actual noise Table 3. Overall accuracies (%) of boosting classifier with no filter, majority vote filtered and margin-based filtered training sets on original data sets 16 Increase in accuracy of 2%

3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Conclusions Ensemble margin = effective and efficient guideline to ensemble design Ensemble learning and ensemble margin are effective for land cover mapping Ensemble margin-based class noise filters are significantly more accurate than majority votes filters. 17