Keypoint-based Recognition Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/04/10.

Slides:



Advertisements
Similar presentations
Kapitel 14 Recognition – p. 1 Recognition Scene understanding / visual object categorization Pose clustering Object recognition by local features Image.
Advertisements

Kapitel 14 Recognition Scene understanding / visual object categorization Pose clustering Object recognition by local features Image categorization Bag-of-features.
Three things everyone should know to improve object retrieval
Presented by Xinyu Chang
Instance-level recognition II.
Feature Matching and Robust Fitting Computer Vision CS 143, Brown James Hays Acknowledgment: Many slides from Derek Hoiem and Grauman&Leibe 2008 AAAI Tutorial.
TP14 - Indexing local features
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Instructor: Mircea Nicolescu Lecture 17
CS4670 / 5670: Computer Vision Bag-of-words models Noah Snavely Object
Photo by Carl Warner.
1 Image Retrieval Hao Jiang Computer Science Department 2009.
Stephan Gammeter, Lukas Bossard, Till Quack, Luc Van Gool.
Image alignment Image from
IBBT – Ugent – Telin – IPI Dimitri Van Cauwelaert A study of the 2D - SIFT algorithm Dimitri Van Cauwelaert.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
CVPR 2008 James Philbin Ondˇrej Chum Michael Isard Josef Sivic
Bag of Features Approach: recent work, using geometric information.
Effective Image Database Search via Dimensionality Reduction Anders Bjorholm Dahl and Henrik Aanæs IEEE Computer Society Conference on Computer Vision.
Robust and large-scale alignment Image from
Object retrieval with large vocabularies and fast spatial matching
1 Image Recognition - I. Global appearance patterns Slides by K. Grauman, B. Leibe.
Lecture 28: Bag-of-words models
Image alignment.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman ICCV 2003 Presented by: Indriyati Atmosukarto.
Distinctive Image Feature from Scale-Invariant KeyPoints
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Sebastian Thrun CS223B Computer Vision, Winter Stanford CS223B Computer Vision, Winter 2005 Lecture 3 Advanced Features Sebastian Thrun, Stanford.
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
Keypoint-based Recognition and Object Search
Alignment and Object Instance Recognition Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/16/12.
Machine learning & category recognition Cordelia Schmid Jakob Verbeek.
10/31/13 Object Recognition and Augmented Reality Computational Photography Derek Hoiem, University of Illinois Dali, Swans Reflecting Elephants.
Object Recognition and Augmented Reality
CS 766: Computer Vision Computer Sciences Department, University of Wisconsin-Madison Indexing and Retrieval James Hill, Ozcan Ilikhan, Mark Lenz {jshill4,
Indexing Techniques Mei-Chen Yeh.
Clustering with Application to Fast Object Search
Image alignment.
CSE 185 Introduction to Computer Vision
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
CSE 473/573 Computer Vision and Image Processing (CVIP)
Fitting and Registration Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/14/12.
Object Tracking/Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition.
04/30/13 Last class: summary, goggles, ices Discrete Structures (CS 173) Derek Hoiem, University of Illinois 1 Image: wordpress.com/2011/11/22/lig.
Lecture 4: Feature matching CS4670 / 5670: Computer Vision Noah Snavely.
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
1 Total Recall: Automatic Query Expansion with a Generative Feature Model for Object Retrieval Ondrej Chum, James Philbin, Josef Sivic, Michael Isard and.
Perceptual and Sensory Augmented Computing Visual Object Recognition Tutorial Visual Object Recognition Bastian Leibe & Computer Vision Laboratory ETH.
Example: line fitting. n=2 Model fitting Measure distances.
Video Google: A Text Retrieval Approach to Object Matching in Videos Josef Sivic and Andrew Zisserman.
10/31/13 Object Recognition and Augmented Reality Computational Photography Derek Hoiem, University of Illinois Dali, Swans Reflecting Elephants.
Lecture 8: Feature matching CS6670: Computer Vision Noah Snavely.
Lecture 08 27/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
CSE 185 Introduction to Computer Vision Feature Matching.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Invariant Local Features Image content is transformed into local feature coordinates that are invariant to translation, rotation, scale, and other imaging.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Prof. Alex Berg (Credits to many other folks on individual slides)
SIFT Scale-Invariant Feature Transform David Lowe
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
CS 2770: Computer Vision Feature Matching and Indexing
Nearest-neighbor matching to feature database
Object Recognition and Augmented Reality
By Suren Manvelyan, Crocodile (nile crocodile?) By Suren Manvelyan,
Nearest-neighbor matching to feature database
Brief Review of Recognition + Context
CSE 185 Introduction to Computer Vision
Presentation transcript:

Keypoint-based Recognition Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/04/10

Today HW 1 HW 2 (p3) Recognition

General Process of Object Recognition Specify Object Model Generate Hypotheses Score Hypotheses Resolution

General Process of Object Recognition Example: Keypoint-based Instance Recognition Specify Object Model Generate Hypotheses Score Hypotheses Resolution B1B1 B2B2 B3B3 A1A1 A2A2 A3A3 Last Class

Overview of Keypoint Matching K. Grauman, B. Leibe N pixels e.g. color B1B1 B2B2 B3B3 A1A1 A2A2 A3A3 1. Find a set of distinctive key- points 3. Extract and normalize the region content 2. Define a region around each keypoint 4. Compute a local descriptor from the normalized region 5. Match local descriptors

General Process of Object Recognition Example: Keypoint-based Instance Recognition Specify Object Model Generate Hypotheses Score Hypotheses Resolution B1B1 B2B2 B3B3 A1A1 A2A2 A3A3 Affine Parameters Choose hypothesis with max score above threshold # Inliers Affine-variant point locations This Class

This class Given two images and their keypoints (position, scale, orientation, descriptor) Goal: Verify if they belong to a consistent configuration Local Features, e.g. SIFT Slide credit: David Lowe

Matching Keypoints Want to match keypoints between: 1.Training images containing the object 2.Database images Given descriptor x 0, find two nearest neighbors x 1, x 2 with distances d 1, d 2 x 1 matches x 0 if d 1 /d 2 < 0.8 – This gets rid of 90% false matches, 5% of true matches in Lowe’s study

Affine Object Model Accounts for 3D rotation of a surface under orthographic projection

Affine Object Model Accounts for 3D rotation of a surface under orthographic projection Scaling/skew Translation How many matched points do we need?

Finding the objects 1.Get matched points in database image 2.Get location/scale/orientation using Hough voting – In training, each point has known position/scale/orientation wrt whole object – Bins for x, y, scale, orientation Loose bins (0.25 object length in position, 2x scale, 30 degrees orientation) Vote for two closest bin centers in each direction (16 votes total) 3.Geometric verification – For each bin with at least 3 keypoints – Iterate between least squares fit and checking for inliers and outliers 4.Report object if > T inliers (T is typically 3, can be computed by probabilistic method)

Matched objects

View interpolation Training – Given images of different viewpoints – Cluster similar viewpoints using feature matches – Link features in adjacent views Recognition – Feature matches may be spread over several training viewpoints  Use the known links to “transfer votes” to other viewpoints Slide credit: David Lowe [Lowe01]

Applications Sony Aibo (Evolution Robotics) SIFT usage – Recognize docking station – Communicate with visual cards Other uses – Place recognition – Loop closure in SLAM K. Grauman, B. Leibe14 Slide credit: David Lowe

Location Recognition Slide credit: David Lowe Training [Lowe04]

Fast visual search “Scalable Recognition with a Vocabulary Tree”, Nister and Stewenius, CVPR “Video Google”, Sivic and Zisserman, ICCV 2003

Slide 110,000,000 Images in 5.8 Seconds Slide Credit: Nister

Slide Slide Credit: Nister

Slide Slide Credit: Nister

Slide

Key Ideas Visual Words – Cluster descriptors (e.g., K-means) Inverse document file – Quick lookup of files given keypoints tf-idf: Term Frequency – Inverse Document Frequency # words in document # times word appears in document # documents # documents that contain the word

Recognition with K-tree Following slides by David Nister (CVPR 2006)

Performance

More words is better Improves Retrieval Improves Speed Branch factor

Higher branch factor works better (but slower)

Video Google System 1.Collect all words within query region 2.Inverted file index to find relevant frames 3.Compare word counts 4.Spatial verification Sivic & Zisserman, ICCV 2003 Demo online at : e/index.html e/index.html 49K. Grauman, B. Leibe Query region Retrieved frames

Sampling strategies 50K. Grauman, B. Leibe Image credits: F-F. Li, E. Nowak, J. Sivic Dense, uniformly Sparse, at interest points Randomly Multiple interest operators To find specific, textured objects, sparse sampling from interest points often more reliable. Multiple complementary interest operators offer more image coverage. For object categorization, dense sampling offers better coverage. [See Nowak, Jurie & Triggs, ECCV 2006]

Application: Large-Scale Retrieval K. Grauman, B. Leibe51 [Philbin CVPR’07] Query Results on 5K (demo available for 100K)

Example Applications B. Leibe52 Mobile tourist guide Self-localization Object/building recognition Photo/video augmentation Aachen Cathedral [Quack, Leibe, Van Gool, CIVR’08]

Application: Image Auto-Annotation K. Grauman, B. Leibe53 Left: Wikipedia image Right: closest match from Flickr [Quack CIVR’08] Moulin Rouge Tour Montparnasse Colosseum Viktualienmarkt Maypole Old Town Square (Prague)

Things to remember Object instance recognition – Find keypoints, compute descriptors – Match descriptors – Vote for / fit affine parameters – Return object if # inliers > T Keys to efficiency – Visual words Used for many applications – Inverse document file Used for web-scale search