Handwritten digit recognition

Slides:



Advertisements
Similar presentations
Classification using intersection kernel SVMs is efficient
Advertisements

Introduction to Support Vector Machines (SVM)
ECG Signal processing (2)
Multiclass SVM and Applications in Object Classification
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
An Introduction of Support Vector Machine
Classification / Regression Support Vector Machines
Handwritten digit recognition
Linear Classifiers/SVMs
Linear Classifiers (perceptrons)
Data Mining Classification: Alternative Techniques
An Introduction of Support Vector Machine
Support Vector Machines and Kernels Adapted from slides by Tim Oates Cognition, Robotics, and Learning (CORAL) Lab University of Maryland Baltimore County.
Classification using intersection kernel SVMs is efficient Joint work with Subhransu Maji and Alex Berg Jitendra Malik UC Berkeley.
Support Vector Machines
SVM—Support Vector Machines
Support vector machine
Machine learning continued Image source:
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Thesis title: “Studies in Pattern Classification – Biological Modeling, Uncertainty Reasoning, and Statistical Learning” 3 parts: (1)Handwritten Digit.
Classification and Decision Boundaries
Fast intersection kernel SVMs for Realtime Object Detection
Discriminative and generative methods for bags of features
Support Vector Machines
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
1 Classification: Definition Given a collection of records (training set ) Each record contains a set of attributes, one of the attributes is the class.
Support Vector Classification (Linearly Separable Case, Primal) The hyperplanethat solves the minimization problem: realizes the maximal margin hyperplane.
Classification Problem 2-Category Linearly Separable Case A- A+ Malignant Benign.
Support Vector Machines and Kernel Methods
Support Vector Machines
K-means Based Unsupervised Feature Learning for Image Recognition Ling Zheng.
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
An Introduction to Support Vector Machines Martin Law.
Classification III Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart Russell,
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Step 3: Classification Learn a decision rule (classifier) assigning bag-of-features representations of images to different classes Decision boundary Zebra.
This week: overview on pattern recognition (related to machine learning)
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 24 – Classifiers 1.
Handwritten digit recognition Jitendra Malik. Handwritten digit recognition (MNIST,USPS) LeCun’s Convolutional Neural Networks variations (0.8%, 0.6%
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
An Introduction to Support Vector Machines (M. Law)
Support Vector Machines Project מגישים : גיל טל ואורן אגם מנחה : מיקי אלעד נובמבר 1999 הטכניון מכון טכנולוגי לישראל הפקולטה להנדסת חשמל המעבדה לעיבוד וניתוח.
Methods for classification and image representation
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
An Introduction to Support Vector Machine (SVM)
SVM – Support Vector Machines Presented By: Bella Specktor.
CS 1699: Intro to Computer Vision Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh October 29, 2015.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Support-Vector Networks C Cortes and V Vapnik (Tue) Computational Models of Intelligence Joon Shik Kim.
Support Vector Machines Optimization objective Machine Learning.
An Introduction of Support Vector Machine In part from of Jinwei Gu.
Finding Clusters within a Class to Improve Classification Accuracy Literature Survey Yong Jae Lee 3/6/08.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Neural networks and support vector machines
Support Vector Machines and Kernels
Recognition using Nearest Neighbor (or kNN)
Digit Recognition using SVMS
Non-linear classifiers Neural networks
An Introduction to Support Vector Machines
Support Vector Machines
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
CS 2750: Machine Learning Support Vector Machines
COSC 4335: Other Classification Techniques
Machine Learning Week 3.
Support Vector Machines and Kernels
Other Classification Models: Support Vector Machine (SVM)
Presentation transcript:

Handwritten digit recognition Jitendra Malik

Handwritten digit recognition (MNIST,USPS) LeCun’s Convolutional Neural Networks variations (0.8%, 0.6% and 0.4% on MNIST) Tangent Distance(Simard, LeCun & Denker: 2.5% on USPS) Randomized Decision Trees (Amit, Geman & Wilder, 0.8%) K-NN based Shape context/TPS matching (Belongie, Malik & Puzicha: 0.6% on MNIST) SVM on orientation histograms(Maji & Malik, 0.8%)

The MNIST DATABASE of handwritten digits yann. lecun The MNIST DATABASE of handwritten digits yann.lecun.com/exdb/mnist/ Yann LeCun & Corinna Cortes Has a training set of 60 K examples (6K examples for each digit), and a test set of 10K examples. Each digit is a 28 x 28 pixel grey level image. The digit itself occupies the central 20 x 20 pixels, and the center of mass lies at the center of the box. “It is a good database for people who want to try learning techniques and pattern recognition methods on real-world data while spending minimal efforts on preprocessing and formatting.”

The machine learning approach to object recognition Training time Compute feature vectors for positive and negative examples of image patches Train a classifier Test Time Compute feature vector on image patch Evaluate classifier

Let us take an example…

Let us take an example…

In feature space, positive and negative examples are just points…

How do we classify a new point?

Nearest neighbor rule “transfer label of nearest example”

Linear classifier rule

Different approaches to training classifiers Nearest neighbor methods Neural networks Support vector machines Randomized decision trees …

Training Stage (+ examples) (- examples) Feature Representation Discriminative Classifier

The multi-scale sliding window approach to detection Ask this question over and over again, varying position, scale, multiple categories…

The multi-scale sliding window approach to detection Ask this question over and over again, varying position, scale, multiple categories…

Support Vector Machines Linear Separators (aka. Perceptrons)

Support Vector Machines Other possible solutions

Support Vector Machines Which one is better? B1 or B2? How do you define better?

Support Vector Machines Find hyperplane maximizes the margin => B1 is better than B2

Support Vector Machines Examples are; (x1,..,xn,y) with y{-1.1}

Some remarks.. While the diagram corresponds to a linearly separable case, the idea can be generalized to a “soft margin SVM” where mistakes are allowed but penalized. Training an SVM is a convex optimization problem, so we are guaranteed that we can find the globally best solution. Various software packages are available such as LIBSVM, LIBLINEAR But what if the decision boundary is horribly non-linear? We use the “kernel trick”

Suppose the positive examples lie inside a disk

Suppose the positive examples lie inside a disk

We can construct a new higher-dimensional feature space where the boundary is linear

Kernel Support Vector Machines Inner Product in Hilbert Space Can Learn Non Linear Boundaries

linear SVM, Kernelized SVM Decision function is where: Linear: Non-linear Using Kernel