Probabilistic Object Recognition and Localization Bernt Schiele, Alex Pentland, ICCV 99 Presenter: Matt Grimes.

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints
Advertisements

Applications of one-class classification
Object Recognition Using Locality-Sensitive Hashing of Shape Contexts Andrea Frome, Jitendra Malik Presented by Ilias Apostolopoulos.
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Object Recognition from Local Scale-Invariant Features David G. Lowe Presented by Ashley L. Kapron.
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Fitting: The Hough transform. Voting schemes Let each feature vote for all the models that are compatible with it Hopefully the noise features will not.
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Mapping: Scaling Rotation Translation Warp
Matching with Invariant Features
Fitting: The Hough transform
1 Interest Operators Find “interesting” pieces of the image –e.g. corners, salient regions –Focus attention of algorithms –Speed up computation Many possible.
A Study of Approaches for Object Recognition
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
1 Interest Operator Lectures lecture topics –Interest points 1 (Linda) interest points, descriptors, Harris corners, correlation matching –Interest points.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2005 with a lot of slides stolen from Steve Seitz and.
Distinctive Image Feature from Scale-Invariant KeyPoints
Distinctive image features from scale-invariant keypoints. David G. Lowe, Int. Journal of Computer Vision, 60, 2 (2004), pp Presented by: Shalomi.
Stepan Obdrzalek Jirı Matas
Object Recognition Using Distinctive Image Feature From Scale-Invariant Key point D. Lowe, IJCV 2004 Presenting – Anat Kaspi.
Fitting a Model to Data Reading: 15.1,
Object Recognition Using Geometric Hashing
Scale Invariant Feature Transform (SIFT)
1 Invariant Local Feature for Object Recognition Presented by Wyman 2/05/2006.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Fitting: The Hough transform
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe – IJCV 2004 Brien Flewelling CPSC 643 Presentation 1.
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
1 Interest Operators Find “interesting” pieces of the image Multiple possible uses –image matching stereo pairs tracking in videos creating panoramas –object.
Overview Introduction to local features
CSE 185 Introduction to Computer Vision
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
Computer vision.
EADS DS / SDC LTIS Page 1 7 th CNES/DLR Workshop on Information Extraction and Scene Understanding for Meter Resolution Image – 29/03/07 - Oberpfaffenhofen.
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
Recognition and Matching based on local invariant features Cordelia Schmid INRIA, Grenoble David Lowe Univ. of British Columbia.
Object Tracking/Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition.
Overview Harris interest points Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points Evaluation and comparison of different.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Fitting: The Hough transform. Voting schemes Let each feature vote for all the models that are compatible with it Hopefully the noise features will not.
Evaluation of interest points and descriptors. Introduction Quantitative evaluation of interest point detectors –points / regions at the same relative.
MSRI workshop, January 2005 Object Recognition Collected databases of objects on uniform background (no occlusions, no clutter) Mostly focus on viewpoint.
Fitting: The Hough transform
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Distinctive Image Features from Scale-Invariant Keypoints Ronnie Bajwa Sameer Pawar * * Adapted from slides found online by Michael Kowalski, Lehigh University.
EECS 274 Computer Vision Model Fitting. Fitting Choose a parametric object/some objects to represent a set of points Three main questions: –what object.
Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison.
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
Local features: detection and description
Distinctive Image Features from Scale-Invariant Keypoints
776 Computer Vision Jan-Michael Frahm Spring 2012.
Hough Transform CS 691 E Spring Outline Hough transform Homography Reading: FP Chapter 15.1 (text) Some slides from Lazebnik.
SIFT.
SIFT Scale-Invariant Feature Transform David Lowe
Presented by David Lee 3/20/2006
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
TP12 - Local features: detection and description
Fitting: The Hough transform
Paper Presentation: Shape and Matching
Feature description and matching
CSE 455 – Guest Lectures 3 lectures Contact Interest points 1
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
SIFT.
Feature descriptors and matching
Presented by Xu Miao April 20, 2005
Recognition and Matching based on local invariant features
Presentation transcript:

Probabilistic Object Recognition and Localization Bernt Schiele, Alex Pentland, ICCV 99 Presenter: Matt Grimes

What they did 1.Chose a set of local image descriptors whose outputs are robust to object orientation and lighting. –Examples: Laplacian First-derivative magnitude:

What they did 2.Learn a PDF for the outputs of these descriptors given an image of the object: Vector of descriptor outputs A particular object Object orientation, lighting, etc.

What they did 2.Learn a PDF for the outputs of these descriptors given an image of the object: Vector of descriptor outputs A particular object

Use Bayes rule to obtain the posterior… …which is the probability of an image containing an object, given local image measurements M. (Not quite this clean) What they did

History of image-based object recognition Two major genres: 1.Histogram-based approaches. 2.Comparison of local image features.

Histogramming approaches Object recognition by color histograms (Swain & Ballard, IJCV 1991) –Robust to changes in orientation, scale. –Brittle against lighting changes (dependency on color). –Many classes of objects not distinguishable by color distribution alone.

Histogramming approaches Combat color-brittleness using (quasi-) invariants of color histograms: –Eigenvalues of matrices of moments of color histograms –Derivatives of logs of color channels –Comprehensive color normalization

Histogramming approaches Comprehensive color normalization examples:

Histogramming approaches Comprehensive color normalization examples:

Localized feature approaches Approaches include: –Using image interest-points to index into a hashtable of known objects. –Comparing large vectors of local filter responses.

Geometric Hashing 1.An interest point detector finds the same points on an object in different images. Types of interest points include corners, T- junctions, sudden texture changes.

Geometric Hashing From Schmid, Mohr, Bauckhage, Comparing and Evaluating Interest Points, ICCV 98

Geometric Hashing From Schmid, Mohr, Bauckhage, Comparing and Evaluating Interest Points, ICCV 98

Geometric Hashing 2.Store points in an affine-transform- invariant representation. 3.Store all possible triplets of points as keys in a hashtable.

Geometric Hashing 4.For object recognition, find all triplets of interest points in an image, look for matches in the hashtable, accumulate votes for the correct object. Hashtable approaches support multiple object recognition within the same image.

Geometric hashing weaknesses Dependent on the consistency of the interest point detector used. From Schmid, Mohr, Bauckhage, Comparing and Evaluating Interest Points, ICCV 98

Geometric hashing weaknesses Shoddy repeatibility necessitates lots of points. Lots of points, combined with noise, leads to lots of false positives.

Vectors of filter responses Typically use vectors of oriented filters at fixed grid points, or at interest points. Pros: –Very robust to noise. Cons: –Fixed grid needs large representation, large grid is sensitive to occlusion. –If using an interest point detector instead, the detector must be consistent over a variety of scenes.

Also: eigenpictures Calculate the eigenpictures of a set of images of objects to be recognized. Pros: –Efficient representation of images by their eigenpicture coefficients. (Fast searches) Cons: –Images must be pre-segmented. –Eigenpictures are not local (sensitive to occlusion). –Translation, image-plane rotation must be represented in the eigenpictures.

This paper: Uses vectors of filter responses, with probabilistic object recognition. Bayes rule Learned from training images Using scene- invariant M

Wins of this paper Uses hashtables for multiple object recognition. Unlike geometric hashing, doesnt depend on point correspondence betw. images. –Uses location-unspecific filter responses, not points. –Inherits robustness to noise of filter response methods.

Wins of this paper Uses local filter responses. –Robust to occlusion compared to global methods (e.g. eigenpictures or filter grids.) Probabilistic matching –Theoretically cleaner than voting. –Combined with local filter responses, allows for localization of detected objects.

Details of the PDF What degrees of freedom are there in the other parameters? o n : Object R: Rotation (3 DOF) T: Translation(3 DOF) S: Scene (occlusions, background) L: Lighting I: Imaging (noise, pixelation/blur)

P(M|o n,R,T,S,L,I) Way too many params to get a reliable estimate from even a large image library. # of examples needed is exponential in the number of dimensions of the PDF. Solution: choose measurements (M) that are invariant with respect to as many params as possible (except o n ).

Techniques for invariance Imaging (noise:) see Schieles thesis. Lighting: apply a energy normalization technique to the filter outputs. Scene: probabilistic object recognition + local image measurements. –Gives best estimate using the visible portion of the object.

Techniques for invariance Translation: –Tx, Ty (image-plane translation) are ignored for non-localizing recognition. –Tz is equivalent to scale. For known scales, compensate by scaling the filters regions of support.

Techniques for invariance Fairly robust to unknown scale:

Techniques for invariance Rotation: –Rz: rotation in the image plane. Filters invariant to image-plane rotation may be used. –Rx, Ry must remain in the PDF. Impossible to have viewpoint- invariant descriptors in the general case.

4 parameters. Still a large amount of training examples needed, but feasible. Example: algorithm has been successful after training with 108 images per object. (108 = 16 orientations * 6 scales) New PDF

Learning & representation of the PDF Since the goal is discrimination, overgeneralization is scarier than overfitting. They chose multidimensional histograms over parametric representations. They mention that they couldve used kernel function estimates.

Multidimensional Histograms

In their experiments, they use a 6-dimensional histogram. –X and Y derivative, at 3 different scales …with 24 buckets per axis. –Theoretical max for # of cells: 24 6 =1.9 x 10 8 Way too many cells to be meaningfully filled by even 512 x 512 (= ) pixel images.

Multidimensional Histograms Somehow, by exploiting dependencies betw. histogram axes, and applying a uniform prior bias, they get the number of calculated cells below Factor of 1000 reduction. Anybody know how they do this?

(Single) object recognition

A single measurement vector m k is insufficient for recognition.

(Single) object recognition A single measurement vector m k is insufficient for recognition.

(Single) object recognition For k measurement vectors:

(Single) object recognition

Measurement regions covering 10~20% of an object are usually sufficient for discrimination.

(Single) object recognition

Multiple object recognition We can apply the single-object detector to many small regions in the image.

Multiple object recognition The algorithm is now O(NKJ) –N = # of known objects –K = # of measurement vectors in each region –J = # of regions

Multiple object recognition

One drawback: For a given image, the algorithm calculates a probability for each object it knows of. The algorithm lists the objects in its library in decreasing order of probability. Need to know beforehand the number of objects in a test image, to know where to stop reading the list.

Failure example

Unfamiliar clutter

Bite the dimensionality bullet and add an object position variable to the PDF: Object localization

Stop assuming independence of mks, to account for structural dependencies: Object localization

Tradeoff between recognition and localization, depending on region size.

Object localization Heirarchical discrimination with coarse fine region size refinement: