Boosting Nearest-Neighbor Classifier for Character Recognition

Slides:



Advertisements
Similar presentations
DECISION TREES. Decision trees  One possible representation for hypotheses.
Advertisements

Ensemble Learning Reading: R. Schapire, A brief introduction to boosting.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
On-line learning and Boosting
Face detection Many slides adapted from P. Viola.
Discriminative and generative methods for bags of features
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
2D1431 Machine Learning Boosting.
Learning of Pseudo-Metrics. Slide 1 Online and Batch Learning of Pseudo-Metrics Shai Shalev-Shwartz Hebrew University, Jerusalem Joint work with Yoram.
A Brief Introduction to Adaboost
Robust supervised image classifiers by spatial AdaBoost based on robust loss functions Ryuei Nishii and Shinto Eguchi Proc. Of SPIE Vol D-2.
Ensemble Learning: An Introduction
Non Parametric Classifiers Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Adaboost and its application
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.
1  The goal is to estimate the error probability of the designed classification system  Error Counting Technique  Let classes  Let data points in class.
Chapter 10 Boosting May 6, Outline Adaboost Ensemble point-view of Boosting Boosting Trees Supervised Learning Methods.
CSSE463: Image Recognition Day 27 This week This week Last night: k-means lab due. Last night: k-means lab due. Today: Classification by “boosting” Today:
A speech about Boosting Presenter: Roberto Valenti.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Benk Erika Kelemen Zsolt
Boosting of classifiers Ata Kaban. Motivation & beginnings Suppose we have a learning algorithm that is guaranteed with high probability to be slightly.
Dünaamiliste süsteemide modelleerimine Identification for control in a non- linear system world Eduard Petlenkov, Automaatikainstituut, 2013.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
1 Ensembles An ensemble is a set of classifiers whose combined results give the final decision. test feature vector classifier 1classifier 2classifier.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Boosting Training Scheme for Acoustic Modeling Rong Zhang and Alexander I. Rudnicky Language Technologies Institute, School of Computer Science Carnegie.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Chapter 4: Pattern Recognition. Classification is a process that assigns a label to an object according to some representation of the object’s properties.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. A fast nearest neighbor classifier based on self-organizing incremental neural network (SOINN) Neuron.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall 2004 AdaBoost.. Binary Classification. Read 9.5 Duda,
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
CSSE463: Image Recognition Day 33 This week This week Today: Classification by “boosting” Today: Classification by “boosting” Yoav Freund and Robert Schapire.
CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect.
Feature Selection and Weighting using Genetic Algorithm for Off-line Character Recognition Systems Faten Hussein Presented by The University of British.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
1 Introduction to Predictive Learning Electrical and Computer Engineering LECTURE SET 8 Combining Methods and Ensemble Learning.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
Face Detection and Recognition Reading: Chapter and, optionally, “Face Recognition using Eigenfaces” by M. Turk and A. Pentland.
AdaBoost Algorithm and its Application on Object Detection Fayin Li.
Adaboost (Adaptive boosting) Jo Yeong-Jun Schapire, Robert E., and Yoram Singer. "Improved boosting algorithms using confidence- rated predictions."
1 Ensembles An ensemble is a set of classifiers whose combined results give the final decision. test feature vector classifier 1classifier 2classifier.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning: Ensemble Methods
Data Mining Practical Machine Learning Tools and Techniques
Handwritten Digit Recognition Using Stacked Autoencoders
Data Mining Practical Machine Learning Tools and Techniques
Week 2 Presentation: Project 3
2. Skin - color filtering.
Advanced data mining with TagHelper and Weka
Boosting and Additive Trees
Importance Weighted Active Learning
Pawan Lingras and Cory Butz
Data Mining Practical Machine Learning Tools and Techniques
Motivation & Introduction Efficiency
CSCI B609: “Foundations of Data Science”
Introduction to Data Mining, 2nd Edition
Introduction to Boosting
network of simple neuron-like computing elements
Presented by: Chang Jia As for: Pattern Recognition
A task of induction to find patterns
Xiao-Yu Zhang, Shupeng Wang, Xiaochun Yun
SPECIAL ISSUE on Document Analysis, 5(2):1-15, 2005.
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Presentation transcript:

Boosting Nearest-Neighbor Classifier for Character Recognition Li Zhang and Lin Liao May 17, 2004

Outline Data set Nearest-Neighbor classifier Boosting algorithm Results

Data set -- MNIST database grayscale handwritten digits All digits are normalized and centered in a 28x28 image Training set: 60,000 examples Test set: 10,000 examples

Nearest-Neighbor Classifier Training Set Test Sample

Motivation of Boosting Nearest-Neighbor Classifier (NN) Need large training set to get good accuracy Classification is slow Boosted NN: Speed up classification by reducing the examples used during classification Do not increase the error rate

Boosting Maintain vector of weights for training samples At each iteration Learn a weak classifier using weighted samples Increase the weights of misclassified examples Combine weak classifiers by weighted voting

Boosting (cont’d) Boost 2-class classifier AdaBoost.M1 Boost multi-class classifier AdaBoost.M2 image label (image, label) [0,1]

AdaBoost.M1

AdaBoost.M1 (cont’d)

AdaBoost.M2

AdaBoost.M2 (cont’d)

Weak Classifier of NN Given the distribution on pairs (X,Y), find a subset P of the training examples and a mapping π: (P,Y) -> {0,1} that achieve small “pseudo-loss”. (|P|=100) Initially P is empty Sample 10 candidates from the distribution (X,Y) Pick 1 candidate from the 10 that minimize the pseudo-loss if it is added to P Repeat 2, 3 until |P|=100 Find the best mapping π given P Return the P, π and the final pseudo-loss

Error Rates

Misclassified Samples 4 / 9 7 / 4 9 / 8 4 / 2 8 / 9 2 / 7 5 / 3 3 / 7 5 / 3 6 / 0 4 / 9 9 / 8 8 / 2 5 / 3 8 / 5

Conclusion Advantages of boosting Disadvantages of boosting The training and test error rates are both theoretically bounded Less “overfitting” in practice Many algorithms can be boosted Easy to implement Disadvantages of boosting Learning is slow