Various camera configurations Single point of fixation where optical axes intersect Optical axes parallel (fixation at infinity) General case.

Slides:



Advertisements
Similar presentations
Lecture 11: Two-view geometry
Advertisements

3D reconstruction.
Geometry 2: A taste of projective geometry Introduction to Computer Vision Ronen Basri Weizmann Institute of Science.
Two-View Geometry CS Sastry and Yang
Two-view geometry.
Camera Calibration. Issues: what are intrinsic parameters of the camera? what is the camera matrix? (intrinsic+extrinsic) General strategy: view calibration.
Camera calibration and epipolar geometry
Camera Models A camera is a mapping between the 3D world and a 2D image The principal camera of interest is central projection.
Structure from motion.
Geometry of Images Pinhole camera, projection A taste of projective geometry Two view geometry:  Homography  Epipolar geometry, the essential matrix.
Stanford CS223B Computer Vision, Winter 2006 Lecture 5 Stereo I
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
Epipolar Geometry and the Fundamental Matrix F
Lecture 21: Multiple-view geometry and structure from motion
Multiple-view Reconstruction from Points and Lines
3D reconstruction of cameras and structure x i = PX i x’ i = P’X i.
CAU Kiel DAGM 2001-Tutorial on Visual-Geometric 3-D Scene Reconstruction 1 The plan for today Leftovers and from last time Camera matrix Part A) Notation,
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
Lecture 20: Two-view geometry CS6670: Computer Vision Noah Snavely.
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
Projected image of a cube. Classical Calibration.
Lec 21: Fundamental Matrix
Structure Computation. How to compute the position of a point in 3- space given its image in two views and the camera matrices of those two views Use.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
Automatic Camera Calibration
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Multi-view geometry.
Geometry and Algebra of Multiple Views
Epipolar geometry The fundamental matrix and the tensor
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
Projective cameras Motivation Elements of Projective Geometry Projective structure from motion Planches : –
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
Projective geometry ECE 847: Digital Image Processing Stan Birchfield Clemson University.
Structure from Motion Course web page: vision.cis.udel.edu/~cv April 25, 2003  Lecture 26.
Multiview Geometry and Stereopsis. Inputs: two images of a scene (taken from 2 viewpoints). Output: Depth map. Inputs: multiple images of a scene. Output:
Day x: Infinity DLT alg HZ 4.1 Rectification HZ 2.7 Hierarchy of maps Invariants HZ 2.4 Projective transform HZ 2.3 Behaviour at infinity Primitives pt/line/conic.
Announcements Project 3 due Thursday by 11:59pm Demos on Friday; signup on CMS Prelim to be distributed in class Friday, due Wednesday by the beginning.
Geometry of Multiple Views
Affine Structure from Motion
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
Two-view geometry. Epipolar Plane – plane containing baseline (1D family) Epipoles = intersections of baseline with image planes = projections of the.
EECS 274 Computer Vision Affine Structure from Motion.
stereo Outline : Remind class of 3d geometry Introduction
Feature Matching. Feature Space Outlier Rejection.
Projective Geometry Hu Zhan Yi. Entities At Infinity The ordinary space in which we lie is Euclidean space. The parallel lines usually do not intersect.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
776 Computer Vision Jan-Michael Frahm & Enrique Dunn Spring 2013.
3D Reconstruction Using Image Sequence
Reconstruction from Two Calibrated Views Two-View Geometry
Stereo March 8, 2007 Suggested Reading: Horn Chapter 13.
Structure from motion Multi-view geometry Affine structure from motion Projective structure from motion Planches : –
EECS 274 Computer Vision Projective Structure from Motion.
Lec 26: Fundamental Matrix CS4670 / 5670: Computer Vision Kavita Bala.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
Parameter estimation class 5
Two-view geometry Computer Vision Spring 2018, Lecture 10
Epipolar geometry.
Computer Graphics Recitation 12.
George Mason University
Two-view geometry.
Two-view geometry.
Multi-view geometry.
Presentation transcript:

Various camera configurations Single point of fixation where optical axes intersect Optical axes parallel (fixation at infinity) General case

Epipolar geometry for cameras in general position

Applications Reconstructing a three dimensional object given two or more photos. The relative orientation of the cameras is unknown, such as when using images downloaded from the internet, taken by different people at different times. Some examples from Univ. Washington/Microsoft Research: – phototour.cs.washington.edu – photosynth.net – grail.cs.washington.edu/rome/

The approach

But first, let us review projective transformations

Projective Transformations Under perspective projection, parallel lines can map to lines that intersect. Therefore, this cannot be modeled by an affine transform! Projective transformations are a more general family which includes affine transforms and perspective projections. Projective transformations are linear transformations using homogeneous coordinates

Homogeneous coordinates Instead of using n coordinates for n- dimensional space, we use n+1 coordinates.

Key rule

Picking a canonical representative

The projective line Any finite point x can be represented as Any infinite point can be expressed as

The projective plane Any finite point can be represented as Any infinite point can be represented as

Lines in homogeneous coordinates

Incidence of points on lines

Line incident on two points

Representing affine transformations

Perspective Projection

Projective transformations

The Essential Matrix Constraint

Proof is based on the co-planarity of three rays

Coplanarity of 3-vectors implies triple product is zero

Longuet-Higgins 8 point algorithm

Summary The basic module of recovering 3D structure from 2 views with relative orientation (R, t) of cameras unknown can be implemented using the Longuet-Higgins 8 point algorithm. The outer loop combines information from all the cameras in a global coordinate system using bundle adjustment. The error that is minimized is the re-projection error. The big idea is that given the guessed 3d positions of a point, one can predict image plane 2d positions in any camera where it is visible. We wish to minimize the squared error between this predicted position and the actual position, summed over all cameras and over all points. Lots of engineering has gone into making these approaches work. Read Szeliski’s book, Chapter 7, for more.