Features-based Object Recognition Pierre Moreels California Institute of Technology Thesis defense, Sept. 24, 2007.

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints
Advertisements

3D Model Matching with Viewpoint-Invariant Patches(VIP) Reporter :鄒嘉恆 Date : 10/06/2009.
Histograms of Oriented Gradients for Human Detection
Human Identity Recognition in Aerial Images Omar Oreifej Ramin Mehran Mubarak Shah CVPR 2010, June Computer Vision Lab of UCF.
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Fitting: The Hough transform. Voting schemes Let each feature vote for all the models that are compatible with it Hopefully the noise features will not.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Recognition by Probabilistic Hypothesis Construction P. Moreels, M. Maire, P. Perona California Institute of Technology.
Image alignment Image from
Ghunhui Gu, Joseph J. Lim, Pablo Arbeláez, Jitendra Malik University of California at Berkeley Berkeley, CA
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Fitting: The Hough transform
Robust and large-scale alignment Image from
Tracking Objects with Dynamics Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 04/21/15 some slides from Amin Sadeghi, Lana Lazebnik,
Real-time Embedded Face Recognition for Smart Home Fei Zuo, Student Member, IEEE, Peter H. N. de With, Senior Member, IEEE.
1 Image Recognition - I. Global appearance patterns Slides by K. Grauman, B. Leibe.
A Study of Approaches for Object Recognition
Generic Object Detection using Feature Maps Oscar Danielsson Stefan Carlsson
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
3-D Object Recognition From Shape Salvador Ruiz Correa Department of Electrical Engineering.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2005 with a lot of slides stolen from Steve Seitz and.
Distinctive Image Feature from Scale-Invariant KeyPoints
Object Recognition Using Distinctive Image Feature From Scale-Invariant Key point D. Lowe, IJCV 2004 Presenting – Anat Kaspi.
Scale Invariant Feature Transform (SIFT)
Evaluation of features detectors and descriptors based on 3D objects P. Moreels - P. Perona California Institute of Technology.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe – IJCV 2004 Brien Flewelling CPSC 643 Presentation 1.
Matthew Brown University of British Columbia (prev.) Microsoft Research [ Collaborators: † Simon Winder, *Gang Hua, † Rick Szeliski † =MS Research, *=MS.
Overview Introduction to local features
Image alignment.
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
Computer vision.
EADS DS / SDC LTIS Page 1 7 th CNES/DLR Workshop on Information Extraction and Scene Understanding for Meter Resolution Image – 29/03/07 - Oberpfaffenhofen.
Recognition and Matching based on local invariant features Cordelia Schmid INRIA, Grenoble David Lowe Univ. of British Columbia.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Building local part models for category-level recognition C. Schmid, INRIA Grenoble Joint work with G. Dorko, S. Lazebnik, J. Ponce.
Fitting: The Hough transform. Voting schemes Let each feature vote for all the models that are compatible with it Hopefully the noise features will not.
1 Total Recall: Automatic Query Expansion with a Generative Feature Model for Object Retrieval Ondrej Chum, James Philbin, Josef Sivic, Michael Isard and.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Features-based Object Recognition P. Moreels, P. Perona California Institute of Technology.
Evaluation of interest points and descriptors. Introduction Quantitative evaluation of interest point detectors –points / regions at the same relative.
CVPR 2003 Tutorial Recognition and Matching Based on Local Invariant Features David Lowe Computer Science Department University of British Columbia.
MSRI workshop, January 2005 Object Recognition Collected databases of objects on uniform background (no occlusions, no clutter) Mostly focus on viewpoint.
Fitting: The Hough transform
Local invariant features Cordelia Schmid INRIA, Grenoble.
Distinctive Image Features from Scale-Invariant Keypoints Ronnie Bajwa Sameer Pawar * * Adapted from slides found online by Michael Kowalski, Lehigh University.
A feature-based kernel for object classification P. Moreels - J-Y Bouguet Intel.
Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison.
CSE 185 Introduction to Computer Vision Feature Matching.
Distinctive Image Features from Scale-Invariant Keypoints
Presented by David Lee 3/20/2006
776 Computer Vision Jan-Michael Frahm Spring 2012.
CSCI 631 – Foundations of Computer Vision March 15, 2016 Ashwini Imran Image Stitching.
Invariant Local Features Image content is transformed into local feature coordinates that are invariant to translation, rotation, scale, and other imaging.
776 Computer Vision Jan-Michael Frahm Spring 2012.
SIFT Scale-Invariant Feature Transform David Lowe
Presented by David Lee 3/20/2006
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Nearest-neighbor matching to feature database
Tracking Objects with Dynamics
Nonparametric Semantic Segmentation
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Nearest-neighbor matching to feature database
CAP 5415 Computer Vision Fall 2012 Dr. Mubarak Shah Lecture-5
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
George Bebis and Wenjing Li Computer Vision Laboratory
CSE 185 Introduction to Computer Vision
Computational Photography
Presented by Xu Miao April 20, 2005
Recognition and Matching based on local invariant features
Presentation transcript:

Features-based Object Recognition Pierre Moreels California Institute of Technology Thesis defense, Sept. 24, 2007

2 The recognition continuum variability Individual objects means of transportation BMW logo Categories cars

Applications Autonomous navigation Identification, Security. Help Daiki find his toys !

4 Problem setup Features Coarse-to-fine algorithm Probabilistic model Experiments Conclusion Outline

5 … The detection problem New scene (test image) Models from database Find models and their pose (location, orientation…)

6 … Hypotheses – models + positions New scene (test image) Models from database 11 22 Θ = affine transformation

7 … Matching features Models from database New scene (test image)  Set of correspondences = assignment vector

8 Features detection

9 Image characterization by features Features = high information content ‘locations in the image where the signal changes two-dimensionally’ C.Schmid Reduce the volume of information edge strength map features –[Sobel 68] –Diff of Gaussians [Crowley84] –[Harris 88] –[Foerstner94] –Entropy [Kadir&Brady01]

10 Correct vs incorrect descriptors matches Mutual Euclidean distances in appearance space of descriptors Pixels intensity within a patch - Steerable filters [Freeman1991] - SIFT [Lowe1999,2004] - Shape context [Belongie2002] - Spin [Johnson1999] - HOG [Dalal2005]

11 Stability with respect to nuisances  Which detector / descriptor combination is best for recognition ?

Past work on evaluation of features Use of flat surfaces, ground truth easily established In 3D images appearance changes more ! [Schmid&Mohr00] [Mikolajczyk&Schmid 03,05,05]

13 Database : 100 3D objects

14 Testing setup [Moreels&Perona ICCV05, IJCV07] Used by [Winder, CVPR07]

Results – viewpoint change Mahalanobis distance No ‘background’ images

2D vs. 3D Ranking of detectors/descriptors combinations are modified when switching from 2D to 3D objects

17 Features matching algorithm

18 Features assignments models from database New scene (test image)... Interpretation...

19 Coarse-to-fine strategy We do it every day ! Search for my place : Los Angeles area – Pasadena – Loma Vista my car

Coarse-to-fine example [Fleuret & Geman 2001,2002] Face identification in complex scenes Coarse resolution Intermediate resolution Fine resolution

21 Progressively narrow down focus on correct region of hypothesis space Reject with little computation cost irrelevant regions of search space Use first information that is easy to obtain Simple building blocks organized in a cascade Probabilistic interpretation of each step Coarse-to-Fine detection

22 Coarse data : prior knowledge Which objects are likely to be there, which pose are they likely to have ? unlikely situations

23 New scene (test image) … Models from database 4 votes 2 votes 0 vote Model voting Search tree (appearance space – leaves = database features)

24 (x 1,y 1,s 1,  1 ) (x 2,y 2,s 2,  2 ) Transform predicted by this match:  x = x 2 -x 1  y = y 2 -y 1  s = s 2 / s 1  =  2 -  1 Each match is represented by a dot in the space of 2D similarities (Hough space) xx yy ss  Use of rich geometric information [Lowe1999,2004]

Prediction of position of model center after transform The space of transform parameters is discretized into ‘bins’ Coarse bins to limit boundary issues and have a low false- alarm rate for this stage We count the number of votes collected by each bin. Coarse Hough transform Model Test scene correct transformation

26 Output of PROSAC : pose transformation + set of features correspondences Correspondence or clutter ? PROSAC Similar to RANSAC – robust statistic for parameter estimation Priority to candidates with good quality of appearance match 2D affine transform : 6 parameters  each sample contains 3 candidate correspondences. d d d [Fischler 1973] [Chum&Matas 2005]

27 Probabilistic model

28 Generative model

29 Recognition steps

Score of an extended hypothesis Hypothesis: model + position observed features geometry + appearance database of models constant Consistency (after PROSAC) Prior on model and poses Features assignments Votes per model Votes per model pose bin (Hough transform) Prior on assignments (before actual observations)

Consistency Consistency between observations and predictions from hypothesis model m position of model m Common-frame approximation : parts are conditionally independent once reference position of the object is fixed. [Lowe1999,Huttenlocher90,Moreels04] Constellation model Common-frame

32 foreground features ‘null’ assignments geometry appearance Consistency - appearanceConsistency - geometry Consistency Consistency between observations and predictions from hypothesis

Learning foreground & background densities Ground truth pairs of matches are collected Gaussian densities, centered on the nomimal value that appearance / pose should have according to H Learning background densities is easy: match to random images. [Moreels&Perona, IJCV, 2007]

34 Experiments

An example Model voting Hough bins

36 An example After PROSAC Probabilistic scores

37 Efficiency of coarse-to-fine processing

38 Giuseppe Toys database – Models 61 objects, 1-2 views/object

Giuseppe Toys database – Test scenes 141 test scenes

40 Home objects database – Models 49 objects, 1-2 views/object

41 Home objects database – Test scenes 141 test scenes

42 Results – Giuseppe Toys database Lowe’99,’04 Lower false alarm rate - more systematic verification of geometry consistency - more consistent verification of geometric consistency undetected objects: features with poor appearance distinctiveness index to incorrect models - +

43 Results – Home objects database

44 Failure mode Test image hand-labeled before the experiments

45 Test – Text and graphics

46 Test – no texture

Test – Clutter

48 Coarse-to-fine strategy prunes irrelevant search branches at early stages. Probabilistic interpretation of each step. Higher performance than Lowe, especially in cluttered environment. Front end (features) needs more work for smooth or shiny surfaces. Conclusions