On Feature Combination for Multiclass Object Classification Peter Gehler and Sebastian Nowozin Reading group October 15, 2009.

Slides:



Advertisements
Similar presentations
Lecture 9 Support Vector Machines
Advertisements

ECG Signal processing (2)
Ensemble Learning Reading: R. Schapire, A brief introduction to boosting.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
Linear Classifiers (perceptrons)
ICML Linear Programming Boosting for Uneven Datasets Jurij Leskovec, Jožef Stefan Institute, Slovenia John Shawe-Taylor, Royal Holloway University.
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
CHAPTER 10: Linear Discrimination
An Introduction of Support Vector Machine
Classification using intersection kernel SVMs is efficient Joint work with Subhransu Maji and Alex Berg Jitendra Malik UC Berkeley.
SVM—Support Vector Machines
Machine learning continued Image source:
Transductive Reliability Estimation for Kernel Based Classifiers 1 Department of Computer Science, University of Ioannina, Greece 2 Faculty of Computer.
Global spatial layout: spatial pyramid matching Spatial weighting the features Beyond bags of features: Adding spatial information.
Second order cone programming approaches for handing missing and uncertain data P. K. Shivaswamy, C. Bhattacharyya and A. J. Smola Discussion led by Qi.
Discriminative and generative methods for bags of features
Support Vector Machine
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Text Classification With Support Vector Machines
Reduced Support Vector Machine
Ensemble Learning: An Introduction
Sketched Derivation of error bound using VC-dimension (1) Bound our usual PAC expression by the probability that an algorithm has 0 error on the training.
CSSE463: Image Recognition Day 31 Due tomorrow night – Project plan Due tomorrow night – Project plan Evidence that you’ve tried something and what specifically.
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
A Kernel-based Support Vector Machine by Peter Axelberg and Johan Löfhede.
What is Learning All about ?  Get knowledge of by study, experience, or being taught  Become aware by information or from observation  Commit to memory.
Support Vector Machines Piyush Kumar. Perceptrons revisited Class 1 : (+1) Class 2 : (-1) Is this unique?
Data mining and machine learning A brief introduction.
CSSE463: Image Recognition Day 27 This week This week Last night: k-means lab due. Last night: k-means lab due. Today: Classification by “boosting” Today:
Support Vector Machines Mei-Chen Yeh 04/20/2010. The Classification Problem Label instances, usually represented by feature vectors, into one of the predefined.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Lecture 6: Classification – Boosting and SVMs CAP 5415 Fall 2006.
Online Multiple Kernel Classification Steven C.H. Hoi, Rong Jin, Peilin Zhao, Tianbao Yang Machine Learning (2013) Presented by Audrey Cheong Electrical.
Boris Babenko, Steve Branson, Serge Belongie University of California, San Diego ICCV 2009, Kyoto, Japan.
CS 1699: Intro to Computer Vision Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh October 29, 2015.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
CSSE463: Image Recognition Day 33 This week This week Today: Classification by “boosting” Today: Classification by “boosting” Yoav Freund and Robert Schapire.
Support Vector Machines (SVM): A Tool for Machine Learning Yixin Chen Ph.D Candidate, CSE 1/10/2002.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Dimensionality reduction
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
1.7 Linear Independence. in R n is said to be linearly independent if has only the trivial solution. in R n is said to be linearly dependent if there.
Notes on HW 1 grading I gave full credit as long as you gave a description, confusion matrix, and working code Many people’s descriptions were quite short.
Support-Vector Networks C Cortes and V Vapnik (Tue) Computational Models of Intelligence Joon Shik Kim.
Hybrid Classifiers for Object Classification with a Rich Background M. Osadchy, D. Keren, and B. Fadida-Specktor, ECCV 2012 Computer Vision and Video Analysis.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
A Parallel Mixture of SVMs for Very Large Scale Problems Ronan Collobert Samy Bengio Yoshua Bengio Prepared : S.Y.C. Neural Information Processing Systems,
CSSE463: Image Recognition Day 14 Lab due Weds. Lab due Weds. These solutions assume that you don't threshold the shapes.ppt image: Shape1: elongation.
Incremental Reduced Support Vector Machines Yuh-Jye Lee, Hung-Yi Lo and Su-Yun Huang National Taiwan University of Science and Technology and Institute.
Introduction to Machine Learning Prof. Nir Ailon Lecture 5: Support Vector Machines (SVM)
LECTURE 20: SUPPORT VECTOR MACHINES PT. 1 April 11, 2016 SDS 293 Machine Learning.
A distributed PSO – SVM hybrid system with feature selection and parameter optimization Cheng-Lung Huang & Jian-Fan Dun Soft Computing 2008.
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
Support Vector Machines
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
An Introduction to Support Vector Machines
A Comparative Study of Kernel Methods for Classification Applications
Vectors, Linear Combinations and Linear Independence
CS 2750: Machine Learning Support Vector Machines
Introduction to Deep Learning with Keras
Multiple Feature Learning for Action Classification
Recitation 6: Kernel SVM
Support Vector Machines
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

On Feature Combination for Multiclass Object Classification Peter Gehler and Sebastian Nowozin Reading group October 15, 2009

Introduction This paper is about: Kernel selection (feature selection) Example: Flower classification  Features: colour and shape  2 kernels  Problem: how to combine these 2 kernels (input to SVM: 1 kernel!)  Simple: take average  Smarter: weighted sum with as many weights as kernels  Even smarter: different weights for each class

Combining kernels – baseline method Compute average over all kernels: Given: distance matrices dl(xi,xj) Goal: compute one single kernel to use with SVMs Recipe:  Compute RBF kernels: kl(xi,xj) = exp(-gl*dl(xi,xj))  Rule-of-thumb: set gl to 1/mean(dl) or 1/median(dl)  Trace normalise each kernel kl such that trace(kl) = 1  Compute average (or product) over all kernels kl

Combining kernels Combination of kernels  Decision function for SVMs: added Multiple Kernel Learning (MKL) Objective function [Varma and Ray] Near identical to l1 C-SVM but added l1 regularisation on the weights d

Combining kernels Combination of kernels  Decision function for SVMs: All kernels share the same alpha and beta values

Combining kernels Boosting of individual kernels Idea:  Learn separate SVMs for each kernel  each with own values for alpha and beta  Use boosting based approach to combine the individual SVMs  linear weighted combination of “weak” classifiers  Authors propose two versions: LP-beta – learns a single weight vector LP-BETA – learns a weight vector for each class

Combining kernels Combination of kernels  Decision function for SVMs:

Results Results on Oxford flowers  7 kernels  Best results when combining multiple kernels  Baseline methods do equally well and are magnitudes faster  The proposed LP methods don’t do better than the baseline either  not explained why!

Results Results on Oxford flowers  adding “noisy” kernels  MKL able to identify these kernels and set weights to ~zero  Accuracy using “averaging” or “product” goes down

Results Results on Caltech-256 dataset  39 kernels  LP-beta performs best  Using the baseline “average” accuracies are within 5% to best results

Results Results on Caltech-101 dataset  LP-beta 10% better than state-of-the-art