Recognition of object by finding correspondences between features of a model and an image. Alignment repeatedly hypothesize correspondences between minimal.

Slides:



Advertisements
Similar presentations
Object Recognition from Local Scale-Invariant Features David G. Lowe Presented by Ashley L. Kapron.
Advertisements

Alignment Visual Recognition “Straighten your paths” Isaiah.
Announcements. Structure-from-Motion Determining the 3-D structure of the world, and/or the motion of a camera using a sequence of images taken by a moving.
Computer vision: models, learning and inference
Two-View Geometry CS Sastry and Yang
Low Complexity Keypoint Recognition and Pose Estimation Vincent Lepetit.
Computer Vision Detecting the existence, pose and position of known objects within an image Michael Horne, Philip Sterne (Supervisor)
Mapping: Scaling Rotation Translation Warp
N-view factorization and bundle adjustment CMPUT 613.
3D M otion D etermination U sing µ IMU A nd V isual T racking 14 May 2010 Centre for Micro and Nano Systems The Chinese University of Hong Kong Supervised.
Camera Calibration. Issues: what are intrinsic parameters of the camera? what is the camera matrix? (intrinsic+extrinsic) General strategy: view calibration.
Camera calibration and epipolar geometry
Image alignment Image from
Structure from motion.
Object Recognition Using Genetic Algorithms CS773C Advanced Machine Intelligence Applications Spring 2008: Object Recognition.
Structure-from-Motion Determining the 3-D structure of the world, and/or the motion of a camera using a sequence of images taken by a moving camera. –Equivalently,
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
Multiple-view Reconstruction from Points and Lines
COMP322/S2000/L221 Relationship between part, camera, and robot (cont’d) the inverse perspective transformation which is dependent on the focal length.
Algebraic Functions of Views for 3D Object Recognition CS773C Advanced Machine Intelligence Applications Spring 2008: Object Recognition.
Fitting a Model to Data Reading: 15.1,
Object Recognition Using Geometric Hashing
Scale Invariant Feature Transform (SIFT)
Recognition by Linear Combinations of Models By Shimon Ullman & Rosen Basri Presented by: Piotr Dollar Nov. 19, 2002.
© 2003 by Davi GeigerComputer Vision October 2003 L1.1 Structure-from-EgoMotion (based on notes from David Jacobs, CS-Maryland) Determining the 3-D structure.
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
COMP322/S2000/L23/L24/L251 Camera Calibration The most general case is that we have no knowledge of the camera parameters, i.e., its orientation, position,
Lecture 8: Image Alignment and RANSAC
The Pinhole Camera Model
Projected image of a cube. Classical Calibration.
MSU CSE 803 Fall 2008 Stockman1 CV: 3D sensing and calibration Coordinate system changes; perspective transformation; Stereo and structured light.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
Camera Calibration CS485/685 Computer Vision Prof. Bebis.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
CSE473/573 – Stereo Correspondence
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
Lecture 10: Robust fitting CS4670: Computer Vision Noah Snavely.
Stockman MSU/CSE Math models 3D to 2D Affine transformations in 3D; Projections 3D to 2D; Derivation of camera matrix form.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
1 Fingerprint Classification sections Fingerprint matching using transformation parameter clustering R. Germain et al, IEEE And Fingerprint Identification.
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
Euclidean cameras and strong (Euclidean) calibration Intrinsic and extrinsic parameters Linear least-squares methods Linear calibration Degenerate point.
Camera Geometry and Calibration Thanks to Martial Hebert.
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
Geometric Models & Camera Calibration
Imaging Geometry for the Pinhole Camera Outline: Motivation |The pinhole camera.
CSCE 643 Computer Vision: Structure from Motion
Affine Structure from Motion
Geometric Hashing: A General and Efficient Model-Based Recognition Scheme Yehezkel Lamdan and Haim J. Wolfson ICCV 1988 Presented by Budi Purnomo Nov 23rd.
EECS 274 Computer Vision Affine Structure from Motion.
Reconnaissance d’objets et vision artificielle Jean Ponce Equipe-projet WILLOW ENS/INRIA/CNRS UMR 8548 Laboratoire.
Structure from Motion ECE 847: Digital Image Processing
776 Computer Vision Jan-Michael Frahm Spring 2012.
Instructor: Mircea Nicolescu Lecture 9
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Lecture 16: Image alignment
Lecture 7: Image alignment
A special case of calibration
Structure from motion Input: Output: (Tomasi and Kanade)
Ellipse Fitting COMP 4900C Winter 2008.
Geometric Hashing: An Overview
Uncalibrated Geometry & Stratification
Course 7 Motion.
Course 6 Stereo.
The Pinhole Camera Model
Calibration and homographies
Structure from motion Input: Output: (Tomasi and Kanade)
Lecture 11: Image alignment, Part 2
Presentation transcript:

Recognition of object by finding correspondences between features of a model and an image. Alignment repeatedly hypothesize correspondences between minimal set of features of a model and an image and then tries to find model poses. For computing poses a model of projection must be selected. A minimal number of points needed to compute a model pose is three. Alignment

General idea: 1.Given an input image and a candidate model, establish correspondence between them. 2.Determine transformation from the model to the image 3.Apply the recovered transformation to the model 4.Compare the transformed model with the viewed object 5.Based on this comparison choose the best model Alignment cont.

General steps before alignment : 1.Selection of object of interest in the picture. 2.Segmentation – delineation of a sub-part of the image to which subsequent recognition process will be applied. 3.Image description – extraction of information which will be used for matching the viewed object with stored object models 4.Extracting an alignment key. Alignment key is an information used to bring the viewed object and models into alignment. Before, During and After Alignment

Alignment 1.Viewed object is brought into correspondence with a large number of models stored in the memory. 2.Individual alignments General steps after alignment: 1.Indexing (classification)– use some criteria to “filter out” unlikely models. Matching Before, During and After Alignment

We consider a work “3D pose from 3 corresponding points under weak perspective projection” of T.D.Alter The problem is to determine the pose of 3 points in space given 3 corresponding points in image. It gives direct expressions for 3 matched model points in image coordinates and an expression of a position in the image of any additional, unmatched model point. 3D pose from 3 corresponding points

Hypothesize a correspondence between three model points and three image points. Compute the 3D pose of the model from three-point correspondence. Predict the image positions of the remaining model points and extended features using the 3D pose. Verify whether the hypothesis is correct by looking in the image near the predicted positions of the model features for corresponding image features. Alignment Algorithm

Fig.1 Model points undergoing perspective projection to produce image points The perspective solution

Let image points be extended as follows: Then The problem is: given find a,b, and c. From the law of cosines: Given a, b, and c, we can compute the 3D locations of the model points: The perspective solution cont.

Approximate perspective projection closely in many cases. Less complicated. Conceptually simpler. We do not need to know the camera focal length and the central point. Fewer solutions (four for perspective an two for weak perspective). Justification of the weak perspective approximation

Weak-Perspective Solution Fig.2 Model points undergoing orthographic projection plus Scale to produce image points

To recover the 3D pose of the model we should know the distances between the model points and distances between the image points The parameters of the geometry in Fig. 2 are (will be proved later ): See eq. (7)-(13). Weak-Perspective Solution cont.

Fig. 3 Small solid representing orthographic projection plus scale of three model points into an image. Computing the Weak-Perspective Solution From Fig. 3 we have three constraints:

Computing the Weak-Perspective Solution cont. Multiplying (3) by –1 and adding all three gives Squaring (4) and using (1) and (2) to eliminate and which leads to biquadratic in s :

where The positive solutions of biquadratic are Computing the Weak-Perspective Solution cont.

From (1),(2) and (4) Computing the Weak-Perspective Solution cont.

The solution fails when the model triangle degenerates to a line, at which case a=0. Computing the Weak-Perspective Solution cont.

Let the image points be Given we can invert the projection to get the tree model points: where unknown w can not be recovered. Image location of a fourth model point.

Denote the model points in arbitrary model coordinate frame. Using solve the following vector equation for the “extended affine coordinates”, of Let Using the three model points with Image location of a fourth model point.

Substituting (17)-(19) into (15) we’ll get Image location of a fourth model point.

To project, first apply the scale factor s : Then project orthographically to get the image location of the fourth point: Image location of a fourth model point.

3D pose from 3 corresponding points under weak perspective projection” T.D.Alter., MIT A.I.Memo No. 1378, D Pose from 3 Points Using Weak Perspective T.D.Alter, IEEE Transactions on Pattern Analysis and Machine Intelligence,v.16 No.8,1994 References