A feature-based kernel for object classification P. Moreels - J-Y Bouguet Intel.

Slides:



Advertisements
Similar presentations
Max-Margin Additive Classifiers for Detection
Advertisements

Distinctive Image Features from Scale-Invariant Keypoints
Shape Matching and Object Recognition using Low Distortion Correspondence Alexander C. Berg, Tamara L. Berg, Jitendra Malik U.C. Berkeley.
Multiclass SVM and Applications in Object Classification
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Three things everyone should know to improve object retrieval
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Linear Classifiers (perceptrons)
MIT CSAIL Vision interfaces Towards efficient matching with random hashing methods… Kristen Grauman Gregory Shakhnarovich Trevor Darrell.
Efficiently searching for similar images (Kristen Grauman)
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
Classification using intersection kernel SVMs is efficient Joint work with Subhransu Maji and Alex Berg Jitendra Malik UC Berkeley.
Machine learning continued Image source:
MIT CSAIL Vision interfaces Approximate Correspondences in High Dimensions Kristen Grauman* Trevor Darrell MIT CSAIL (*) UT Austin…
The Pyramid Match Kernel: Discriminative Classification with Sets of Image Features Kristen Grauman Trevor Darrell MIT.
Learning Visual Similarity Measures for Comparing Never Seen Objects Eric Nowak, Frédéric Jurie CVPR 2007.
1 Part 1: Classical Image Classification Methods Kai Yu Dept. of Media Analytics NEC Laboratories America Andrew Ng Computer Science Dept. Stanford University.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Addressing the Medical Image Annotation Task using visual words representation Uri Avni, Tel Aviv University, Israel Hayit GreenspanTel Aviv University,
Global spatial layout: spatial pyramid matching Spatial weighting the features Beyond bags of features: Adding spatial information.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Fast intersection kernel SVMs for Realtime Object Detection
Discriminative and generative methods for bags of features
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Computational Photography
Recognition using Regions CVPR Outline Introduction Overview of the Approach Experimental Results Conclusion.
Robust and large-scale alignment Image from
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
A Study of Approaches for Object Recognition
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Features-based Object Recognition Pierre Moreels California Institute of Technology Thesis defense, Sept. 24, 2007.
CS294‐43: Visual Object and Activity Recognition Prof. Trevor Darrell Spring 2009 March 17 th, 2009.
1 Diffusion Distance for Histogram Comparison, CVPR06. Haibin Ling, Kazunori Okada Group Meeting Presented by Wyman 3/14/2006.
Distinctive image features from scale-invariant keypoints. David G. Lowe, Int. Journal of Computer Vision, 60, 2 (2004), pp Presented by: Shalomi.
(1) Feature-point matching by D.J.Duff for CompVis Online: Feature Point Matching Detection, Extraction.
Scale Invariant Feature Transform (SIFT)
Local Features and Kernels for Classification of Object Categories J. Zhang --- QMUL UK (INRIA till July 2005) with M. Marszalek and C. Schmid --- INRIA.
1 Invariant Local Feature for Object Recognition Presented by Wyman 2/05/2006.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe – IJCV 2004 Brien Flewelling CPSC 643 Presentation 1.
Object Recognition by Discriminative Methods Sinisa Todorovic 1st Sino-USA Summer School in VLPR July, 2009.
Sebastian Thrun CS223B Computer Vision, Winter Stanford CS223B Computer Vision, Winter 2005 Lecture 3 Advanced Features Sebastian Thrun, Stanford.
Discriminative and generative methods for bags of features
A String Matching Approach for Visual Retrieval and Classification Mei-Chen Yeh* and Kwang-Ting Cheng Learning-Based Multimedia Lab Department of Electrical.
Efficient Algorithms for Matching Pedro Felzenszwalb Trevor Darrell Yann LeCun Alex Berg.
Unsupervised Learning of Categories from Sets of Partially Matching Image Features Kristen Grauman and Trevor Darrel CVPR 2006 Presented By Sovan Biswas.
Computer vision.
Step 3: Classification Learn a decision rule (classifier) assigning bag-of-features representations of images to different classes Decision boundary Zebra.
Problem Statement A pair of images or videos in which one is close to the exact duplicate of the other, but different in conditions related to capture,
Local invariant features Cordelia Schmid INRIA, Grenoble.
Reporter: Fei-Fei Chen. Wide-baseline matching Object recognition Texture recognition Scene classification Robot wandering Motion tracking.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Features-based Object Recognition P. Moreels, P. Perona California Institute of Technology.
In Defense of Nearest-Neighbor Based Image Classification Oren Boiman The Weizmann Institute of Science Rehovot, ISRAEL Eli Shechtman Adobe Systems Inc.
Computer Vision Group University of California Berkeley On Visual Recognition Jitendra Malik UC Berkeley.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
CSE 185 Introduction to Computer Vision Feature Matching.
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
Presented by David Lee 3/20/2006
776 Computer Vision Jan-Michael Frahm Spring 2012.
Invariant Local Features Image content is transformed into local feature coordinates that are invariant to translation, rotation, scale, and other imaging.
SIFT Scale-Invariant Feature Transform David Lowe
Distinctive Image Features from Scale-Invariant Keypoints
Learning Mid-Level Features For Recognition
Paper Presentation: Shape and Matching
Approximate Correspondences in High Dimensions
CS 1674: Intro to Computer Vision Scene Recognition
A Graph-Matching Kernel for Object Categorization
Presentation transcript:

A feature-based kernel for object classification P. Moreels - J-Y Bouguet Intel

One scenario for image query : Relevance Feedback Relevance Feedback ImageDatabase Query + “Baby with house” SimilarityMetricCandidateResults Final Result SVM classifier ???

Outline Motivation Features and distances Starting point: the pyramid match kernel Extension : the max-kernel. Experiments Conclusions

Similarity Metric between pairs of images Need to derive a measure of similarity between images How similar are those two images?

Similarity Metric between pairs of images Need to derive a measure of similarity between images How similar are those two images? It is desired that:

Image descriptors: features Smaller volume of information Robustness to deformations (lighting, rotation, affine transformations) Detector: difference-of-gaussians Descriptor: SIFT = 128D collection of local gradients

Match-based Similarity IDEA: First establish correspondence between the two sets of points and then compute a “distance” metric based on the matching result STEP 1: Establish Correspondence STEP 2: Compute similarity PROBLEM: Until now, no matching-based image similarity metric has been shown to satisfy the MERCER conditions

Our main contribution Derive a new image similarity metric that is based on point correspondence and satisfies the Mercer condition Methodoly: generalize another metric developed by Grauman at MIT (pyramid kernel) while preserving its Mercer quality

Pyramid match – K.Grauman (ICCV05) Only appearance is considered Image represented in terms of multi-scale histograms Appearance space

Matching process Soft matches by histogram intersection Fine resolution to coarse resolution More weight at fine resolution: 2 -level =1/size(bin) Level 0 Level 1 Level 2

Final score (Kernel) count intersection at current level discards matches already counted at previous levels - Fine resolution first - Coarse resolution last This kernel verifies Mercer condition ! (Odone et all, TIP, 2005) more weight given to best matches

Issues should be matched at this level level 0 level 1 level 2 level 3 counted only here Boundary problems x 2

Issues level n 2 level =size(bin) approximates poorly the distance between 2 points Weight function f(d)= 1/d over- emphasizes small distances w = weight = 1/size(bin) c = correct weight = 1/d (w-c)/c

From discrete to continous 2 k is a poor approximation  increase the number of resolution steps Boundary problems  use translation of bins level 0 level 1 level 2 level 3 Verifies Mercer condition  (1+  ) 11 00 22 33

STEP 1: Establish Correspondence Our kernel This kernel is easy to compute Uses exact distances No over-emphasis of low distances still verifies Mercer condition best matches first where

Experiments – distance accuracy Random sets of 2D points Compares distances based on: our kernel, pyramid match, Earth Mover’s Distance (EMD = optimal solution to the matching problem, based on simplex) Distances measured between simulated images for pyramid match, max-kernel and EMD Corresponding probability density function

Caltech database – 101 categories

Some data ~50 to ~300 images per class Performance of the competitors: –Chance : 1% –Fei-Fei & Perona : 16% –Berg & Malik : 48% –Holub & Perona : 40% –Grauman & Darrell : 43% Classification using a SVM

Classification results category vs. bg:  performance = 89% 10 random categories  performance: 61%

Classification results 7 good, 7 bad categories Performance: 45%

Computation time features per image 0.1s to compute kernel value b/w 2 images Training: each classifier (100 classifiers) requires x kernel values SVM optimization is not an issue Testing: most training samples are SVMs  of the order of 30 x 80 kernel values per class Parallelization: everything is computed independently –in the training phase –in the test phase

Conclusions The MaxKernel is more accurate than pyramid match, more practical than EMD Good approximation of the optimal distance Verifies Mercer  SVM classification OK Initial classification performance in same ballpark as the competition TODO: add some geometry – e.g. Hough transform to filter out wrong correspondences