Boris Babenko, Steve Branson, Serge Belongie University of California, San Diego ICCV 2009, Kyoto, Japan.

Slides:



Advertisements
Similar presentations
Ignas Budvytis*, Tae-Kyun Kim*, Roberto Cipolla * - indicates equal contribution Making a Shallow Network Deep: Growing a Tree from Decision Regions of.
Advertisements

Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
On-line learning and Boosting
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Boosting Rong Jin.
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
Efficient Large-Scale Structured Learning
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
Support Vector Machines
Machine learning continued Image source:
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Gustavo Carneiro The Automatic Design of Feature Spaces for Local Image Descriptors using an Ensemble of Non-linear Feature Extractors.
Online Multiple Classifier Boosting for Object Tracking Tae-Kyun Kim 1 Thomas Woodley 1 Björn Stenger 2 Roberto Cipolla 1 1 Dept. of Engineering, University.
CS4670 / 5670: Computer Vision Bag-of-words models Noah Snavely Object
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Global spatial layout: spatial pyramid matching Spatial weighting the features Beyond bags of features: Adding spatial information.
EE462 MLCV Lecture 5-6 Object Detection – Boosting Tae-Kyun Kim.
Contour Based Approaches for Visual Object Recognition Jamie Shotton University of Cambridge Joint work with Roberto Cipolla, Andrew Blake.
Sketch Tokens: A Learned Mid-level Representation for Contour and Object Detection CVPR2013 POSTER.
Discriminative and generative methods for bags of features
Bag-of-features models Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Lecture 28: Bag-of-words models
Bag-of-features models
Clustering.
Supervised Distance Metric Learning Presented at CMU’s Computer Vision Misc-Read Reading Group May 9, 2007 by Tomasz Malisiewicz.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Adapted by Doug Downey from Machine Learning EECS 349, Bryan Pardo Machine Learning Clustering.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Large Scale Recognition and Retrieval. What does the world look like? High level image statistics Object Recognition for large-scale search Focus on scaling.
Review: Intro to recognition Recognition tasks Machine learning approach: training, testing, generalization Example classifiers Nearest neighbor Linear.
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Exercise Session 10 – Image Categorization
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Data mining and machine learning A brief introduction.
DATA MINING CLUSTERING K-Means.
Visual Tracking with Online Multiple Instance Learning
Jifeng Dai 2011/09/27.  Introduction  Structural SVM  Kernel Design  Segmentation and parameter learning  Object Feature Descriptors  Experimental.
Recognition using Boosting Modified from various sources including
Support Vector Machines Mei-Chen Yeh 04/20/2010. The Classification Problem Label instances, usually represented by feature vectors, into one of the predefined.
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
Benk Erika Kelemen Zsolt
Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
Boris Babenko 1, Ming-Hsuan Yang 2, Serge Belongie 1 1. University of California, San Diego 2. University of California, Merced OLCV, Kyoto, Japan.
Boris 2 Boris Babenko 1 Ming-Hsuan Yang 2 Serge Belongie 1 (University of California, Merced, USA) 2 (University of California, San Diego, USA) Visual.
Representations for object class recognition David Lowe Department of Computer Science University of British Columbia Vancouver, Canada Sept. 21, 2006.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Clustering What is clustering? Also called “unsupervised learning”Also called “unsupervised learning”
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Methods for classification and image representation
Associative Hierarchical CRFs for Object Class Image Segmentation
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Ariadna Quattoni Xavier Carreras An Efficient Projection for l 1,∞ Regularization Michael Collins Trevor Darrell MIT CSAIL.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
Notes on HW 1 grading I gave full credit as long as you gave a description, confusion matrix, and working code Many people’s descriptions were quite short.
Lecture 5: Statistical Methods for Classification CAP 5415: Computer Vision Fall 2006.
CS Machine Learning Instance Based Learning (Adapted from various sources)
Learning Kernel Classifiers 1. Introduction Summarized by In-Hee Lee.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Semi-Supervised Clustering
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Metric Learning for Clustering
CS 1674: Intro to Computer Vision Scene Recognition
A Graph-Matching Kernel for Object Categorization
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Boris Babenko, Steve Branson, Serge Belongie
Presentation transcript:

Boris Babenko, Steve Branson, Serge Belongie University of California, San Diego ICCV 2009, Kyoto, Japan

Recognizing multiple categories – Need meaningful similarity metric / feature space

Recognizing multiple categories – Need meaningful similarity metric / feature space Idea: use training data to learn metric, plug into kNN – Goes by many names: metric learning cue combination/weighting kernel combination/learning feature selection

Learn a single global similarity metric Labeled Dataset Monolithic Query Image Similarity Metric Category 4 Category 3 Category 2 Category 1 [ Jones et al. ‘03, Chopra et al. ‘05, Goldberger et al. ‘05, Shakhnarovich et al. ‘05 Torralba et al. ‘08]

Learn similarity metric for each category (1-vs-all) Labeled Dataset Monolithic Category Specific Query Image Similarity Metric Category 4 Category 3 Category 2 Category 1 [ Varma et al. ‘07, Frome et al. ‘07, Weinberger et al. ‘08 Nilsback et al. ’08]

Per category: – More powerful – Do we really need thousands of metrics? – Have to train for new categories Global/Monolithic: – Less powerful – Can generalize to new categories

Would like to explore space between two extremes Idea: – Group categories together – Learn a few similarity metrics, one for each super- category

Learn a few good similarity metrics Labeled Dataset Monolithic MuSL Category Specific Query Image Similarity Metric Category 4 Category 3 Category 2 Category 1

Need some framework to work with… Boosting has many advantages: – Feature selection – Easy implementation – Performs well Can treat metric learning as binary classification

Training data: Generate pairs: – Sample negative pairs (, ), 0 Images Category Labels (, ), 1

Train similarity metric/classifier:

Choose to be binary -- i.e. = L1 distance over binary vectors – Can pre-compute for training data – Efficient to compute (XOR and sum) For convenience: [Shakhnarovich et al. ’05, Fergus et al. ‘08]

Given some objective function Boosting = gradient ascent in function space Gradient = example weights for boosting chosen weak classifier other weak classifiers function space current strong classifier [Friedman ’01, Mason et al. ‘00]

Goal: trainand recover mapping At runtime – To compute similarity of query image to use Category 4 Category 3 Category 2 Category 1

Run pre-processing to group categories (i.e. k- means), then train as usual Drawbacks: – Hacky / not elegant – Not optimal: pre-processing not informed by class confusions, etc. How can we train & group simultaneously?

Definitions: Sigmoid FunctionParameter

Definitions:

How well works with category

Objective function: Each category “assigned” to classifier

Replace max with differentiable approx. where is a scalar parameter

Each training pair has weights

Intuition: Approximation of Difficulty of pair (like regular boosting)

(boosting iteration) Difficult Pair Assigned to Easy Pair Assigned to

for - Compute weights - Train on weighted pairs end Assign

Created dataset with hierarchical structure of categories Merged categories from: Caltech 101 [Griffin et al.] Oxford Flowers [Nilsback et al.] UIUC Textures [Lazebnik et al.]

MuSL k-means

Training more metrics overfits! New categories onlyBoth new and old categories

Studied categorization performance vs number of learned metrics Presented boosting algorithm to simultaneously group categories and train metrics Observed overfitting behavior for novel categories

Supported by – NSF CAREER Grant # – NSF IGERT Grant DGE – ONR MURI Grant #N – UCSD FWGrid Project (NSF Infrastructure Grant no. EIA )

Train similarity metric/classifier: Let then

Matrix: Caltech Flowers Textures Caltech Flowers Textures Caltech Flowers Textures - High value - Low value