Boris Babenko, Steve Branson, Serge Belongie

Slides:



Advertisements
Similar presentations
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Advertisements

On-line learning and Boosting
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
Efficient Large-Scale Structured Learning
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
Machine learning continued Image source:
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Gustavo Carneiro The Automatic Design of Feature Spaces for Local Image Descriptors using an Ensemble of Non-linear Feature Extractors.
Online Multiple Classifier Boosting for Object Tracking Tae-Kyun Kim 1 Thomas Woodley 1 Björn Stenger 2 Roberto Cipolla 1 1 Dept. of Engineering, University.
Ziming Zhang *, Ze-Nian Li, Mark Drew School of Computing Science, Simon Fraser University, Vancouver, B.C., Canada {zza27, li, Learning.
Computer vision: models, learning and inference
CS4670 / 5670: Computer Vision Bag-of-words models Noah Snavely Object
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Global spatial layout: spatial pyramid matching Spatial weighting the features Beyond bags of features: Adding spatial information.
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
Contour Based Approaches for Visual Object Recognition Jamie Shotton University of Cambridge Joint work with Roberto Cipolla, Andrew Blake.
Discriminative and generative methods for bags of features
Bag-of-features models Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Paper Discussion: “Simultaneous Localization and Environmental Mapping with a Sensor Network”, Marinakis et. al. ICRA 2011.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Lecture 28: Bag-of-words models
Graz University of Technology, AUSTRIA Institute for Computer Graphics and Vision Fast Visual Object Identification and Categorization Michael Grabner,
Supervised Distance Metric Learning Presented at CMU’s Computer Vision Misc-Read Reading Group May 9, 2007 by Tomasz Malisiewicz.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Large Scale Recognition and Retrieval. What does the world look like? High level image statistics Object Recognition for large-scale search Focus on scaling.
Review: Intro to recognition Recognition tasks Machine learning approach: training, testing, generalization Example classifiers Nearest neighbor Linear.
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Data mining and machine learning A brief introduction.
Visual Tracking with Online Multiple Instance Learning
Object Detection Sliding Window Based Approach Context Helps
Jifeng Dai 2011/09/27.  Introduction  Structural SVM  Kernel Design  Segmentation and parameter learning  Object Feature Descriptors  Experimental.
Recognition using Boosting Modified from various sources including
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.
Boris Babenko 1, Ming-Hsuan Yang 2, Serge Belongie 1 1. University of California, San Diego 2. University of California, Merced OLCV, Kyoto, Japan.
Boris 2 Boris Babenko 1 Ming-Hsuan Yang 2 Serge Belongie 1 (University of California, Merced, USA) 2 (University of California, San Diego, USA) Visual.
Representations for object class recognition David Lowe Department of Computer Science University of British Columbia Vancouver, Canada Sept. 21, 2006.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Clustering What is clustering? Also called “unsupervised learning”Also called “unsupervised learning”
Robust Object Tracking with Online Multiple Instance Learning
Methods for classification and image representation
Boris Babenko, Steve Branson, Serge Belongie University of California, San Diego ICCV 2009, Kyoto, Japan.
Ariadna Quattoni Xavier Carreras An Efficient Projection for l 1,∞ Regularization Michael Collins Trevor Darrell MIT CSAIL.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Neural networks and support vector machines
Clustering Data Streams
Semi-Supervised Clustering
Deep Feedforward Networks
Artificial Neural Networks
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Data Driven Attributes for Action Detection
Session 7: Face Detection (cont.)
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Geometrical intuition behind the dual problem
Paper Presentation: Shape and Matching
ECE 5424: Introduction to Machine Learning
Asymmetric Gradient Boosting with Application to Spam Filtering
Boosting Nearest-Neighbor Classifier for Character Recognition
A weight-incorporated similarity-based clustering ensemble method based on swarm intelligence Yue Ming NJIT#:
CS 1674: Intro to Computer Vision Scene Recognition
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei.
Multiple Instance Learning: applications to computer vision
Presented by: Chang Jia As for: Pattern Recognition
A Graph-Matching Kernel for Object Categorization
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Reinforcement Learning (2)
Presentation transcript:

Boris Babenko, Steve Branson, Serge Belongie Similarity Metrics for Categorization: From Monolithic to Category Specific Boris Babenko, Steve Branson, Serge Belongie University of California, San Diego ICCV 2009, Kyoto, Japan

Similarity Metrics for Recognition Recognizing multiple categories Need meaningful similarity metric / feature space

Similarity Metrics for Recognition Recognizing multiple categories Need meaningful similarity metric / feature space Idea: use training data to learn metric Goes by many names: metric learning cue combination/weighting kernel combination/learning feature selection

Similarity Metrics for Recognition Learn a single global similarity metric Labeled Dataset Monolithic Query Image Similarity Metric Category 4 Category 3 Category 2 Category 1 [ Jones et al. ‘03, Chopra et al. ‘05, Goldberger et al. ‘05, Shakhnarovich et al. ‘05 Torralba et al. ‘08]

Similarity Metrics for Recognition Learn similarity metric for each category (1-vs-all) Labeled Dataset Monolithic Category Specific Query Image Similarity Metric Category 4 Category 3 Category 2 Category 1 What if the number of categories is 10000… do we need 10000 to get good performance [ Varma et al. ‘07, Frome et al. ‘07, Weinberger et al. ‘08 Nilsback et al. ’08]

How many should we train? Monolithic: Less powerful… there is no “perfect” metric Can generalize to new categories Per category: More powerful Do we really need thousands of metrics? Have to train for new categories

Multiple Similarity Learning (MuSL) Would like to explore space between two extremes Idea: Group categories together Learn a few similarity metrics, one for each group - Some example…

Multiple Similarity Learning (MuSL) Learn a few good similarity metrics Query Image Similarity Metric Labeled Dataset Monolithic Category 1 Category 2 MuSL Category 3 Category Specific Category 4

Review of Boosting Similarity Need some framework to work with… Boosting has many advantages: Feature selection Easy implementation Performs well

Notation Training data: Generate pairs: Sample negative pairs ( , ), 1 Images Category Labels ( , ), 1 ( , ), 0

Boosting Similarity Train similarity metric/classifier:

Boosting Similarity Choose to be binary -- i.e. = L1 distance over binary vectors efficient to compute (XOR and sum) For convenience: [Shakhnarovich et al. ’05, Fergus et al. ‘08]

Gradient Boosting Given some objective function Boosting = gradient ascent in function space Gradient = example weights for boosting chosen weak classifier current strong classifier The way we do this is by interperting boosting as gradient ascent in function space. We would like to find a point in this space that optimizes the objective function. Each weak classifier is a vector in this space, so to build up our strong classifier, we compute the gradient of the objective function, and find a weak classifier as close as possible to this direction. We can think of this gradient as a vector of weights, one for each training example. other weak classifiers function space [Friedman ’01, Mason et al. ‘00]

MuSL Boosting Goal: train and recover mapping At runtime To compute similarity of query image to use Category 4 Category 3 Category 2 Category 1 Add slide about kmeans… not informed by class confusions

Naïve Solution Run pre-processing to group categories (i.e. k-means), then train as usual Drawbacks: Hacky / not elegant Not optimal: pre-processing not informed by class confusions, etc. How can we train & group simultaneously?

MuSL Boosting Definitions: Sigmoid Function Parameter

MuSL Boosting Definitions:

MuSL Boosting Definitions: How well works with category

MuSL Boosting Objective function: Each category “assigned” to classifier

Approximating Max Replace max with differentiable approx. where is a scalar parameter

Pair Weights Each training pair has weights

(like regular boosting) Pair Weights Intuition: Approximation of Difficulty of pair (like regular boosting)

Evolution of Weights Difficult Pair Assigned to Easy Pair Assigned to (boosting iteration) (boosting iteration)

MuSL Boosting Algorithm for - Compute weights - Train on weighted pairs end Assign

MuSL Results Created dataset with many heterogeneous categories Merged categories from: Caltech 101 [Griffin et al.] Oxford Flowers [Nilsback et al.] UIUC Textures [Lazebnik et al.]

Recovered Groupings MuSL k-means

Generalizing to New Categories Training more metrics overfits!

Conclusions Studied categorization performance vs number of learned metrics Presented boosting algorithm to simultaneously group categories and train metrics Observed overfitting behavior for novel categories

Thank you! Supported by NSF CAREER Grant #0448615 NSF IGERT Grant DGE-0333451 ONR MURI Grant #N00014-08-1-0638 UCSD FWGrid Project (NSF Infrastructure Grant no. EIA-0303622)