Calibration.

Slides:



Advertisements
Similar presentations
Single-view geometry Odilon Redon, Cyclops, 1914.
Advertisements

Lecture 11: Two-view geometry
Last 4 lectures Camera Structure HDR Image Filtering Image Transform.
Computer vision: models, learning and inference
Two-View Geometry CS Sastry and Yang
Two-view geometry.
Camera Calibration. Issues: what are intrinsic parameters of the camera? what is the camera matrix? (intrinsic+extrinsic) General strategy: view calibration.
Camera calibration and epipolar geometry
Single-view metrology
3D Geometry and Camera Calibration. 3D Coordinate Systems Right-handed vs. left-handedRight-handed vs. left-handed x yz x yz.
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
Epipolar Geometry and the Fundamental Matrix F
Many slides and illustrations from J. Ponce
Uncalibrated Epipolar - Calibration
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
Calibration Dorit Moshe.
Panorama signups Panorama project issues? Nalwa handout Announcements.
Lecture 16: Single-view modeling, Part 2 CS6670: Computer Vision Noah Snavely.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Single-view geometry Odilon Redon, Cyclops, 1914.
Projected image of a cube. Classical Calibration.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
CS 558 C OMPUTER V ISION Lecture IX: Dimensionality Reduction.
Today: Calibration What are the camera parameters?
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
Projective Geometry and Single View Modeling CSE 455, Winter 2010 January 29, 2010 Ames Room.
776 Computer Vision Jan-Michael Frahm, Enrique Dunn Spring 2013.
Automatic Camera Calibration
Euclidean cameras and strong (Euclidean) calibration Intrinsic and extrinsic parameters Linear least-squares methods Linear calibration Degenerate point.
Camera Geometry and Calibration Thanks to Martial Hebert.
Multi-view geometry.
Epipolar geometry The fundamental matrix and the tensor
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
Projective cameras Motivation Elements of Projective Geometry Projective structure from motion Planches : –
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
CSCE 643 Computer Vision: Structure from Motion
Multiview Geometry and Stereopsis. Inputs: two images of a scene (taken from 2 viewpoints). Output: Depth map. Inputs: multiple images of a scene. Output:
Geometric Camera Models
Vision Review: Image Formation Course web page: September 10, 2002.
Peripheral drift illusion. Multiple views Hartley and Zisserman Lowe stereo vision structure from motion optical flow.
Geometry of Multiple Views
Affine Structure from Motion
Single-view geometry Odilon Redon, Cyclops, 1914.
Camera Calibration Sebastian Thrun, Gary Bradski, Daniel Russakoff Stanford CS223B Computer Vision (with material from.
Ch. 3: Geometric Camera Calibration
CS-498 Computer Vision Week 7, Day 2 Camera Parameters Intrinsic Calibration  Linear  Radial Distortion (Extrinsic Calibration?) 1.
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
EECS 274 Computer Vision Geometric Camera Calibration.
EECS 274 Computer Vision Affine Structure from Motion.
Computer vision: models, learning and inference M Ahad Multiple Cameras
MASKS © 2004 Invitation to 3D vision Uncalibrated Camera Chapter 6 Reconstruction from Two Uncalibrated Views Modified by L A Rønningen Oct 2008.
Single-view geometry Odilon Redon, Cyclops, 1914.
EECS 274 Computer Vision Projective Structure from Motion.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
Calibrating a single camera
Geometric Camera Calibration
Depth from disparity (x´,y´)=(x+D(x,y), y)
Two-view geometry Computer Vision Spring 2018, Lecture 10
Epipolar geometry.
Structure from motion Input: Output: (Tomasi and Kanade)
Multiple View Geometry for Robotics
Uncalibrated Geometry & Stratification
Two-view geometry.
Multi-view geometry.
Single-view geometry Odilon Redon, Cyclops, 1914.
Structure from motion Input: Output: (Tomasi and Kanade)
Presentation transcript:

Calibration

Camera Calibration Geometric Radiometric Intrinsics: Focal length, principal point, distortion Extrinsics: Position, orientation Radiometric Mapping between pixel value and scene radiance Can be nonlinear at a pixel (gamma, etc.) Can vary between pixels (vignetting, cos4, etc.) Dynamic range (calibrate shutter speed, etc.)

Geometric Calibration Issues Camera Model Orthogonal axes? Square pixels? Distortion? Calibration Target Known 3D points, noncoplanar Known 3D points, coplanar Unknown 3D points (structure from motion) Other features (e.g., known straight lines)

Geometric Calibration Issues Optimization method Depends on camera model, available data Linear vs. nonlinear model Closed form vs. iterative Intrinsics only vs. extrinsics only vs. both Need initial guess?

Caveat - 2D Coordinate Systems y axis up vs. y axis down Origin at center vs. corner Will often write (u, v) for image coordinates u v v u v u

Camera Calibration – Example 1 Given: 3D  2D correspondences General perspective camera model (no distortion) Don’t care about “z” after transformation Homogeneous scale ambiguity  11 free parameters

Camera Calibration – Example 1 Write equations:

Camera Calibration – Example 1 Linear equation Overconstrained (more equations than unknowns) Underconstrained (rank deficient matrix – any multiple of a solution, including 0, is also a solution)

Camera Calibration – Example 1 Standard linear least squares methods for Ax=0 will give the solution x=0 Instead, look for a solution with |x|= 1 That is, minimize |Ax|2 subject to |x|2=1

Camera Calibration – Example 1 Minimize |Ax|2 subject to |x|2=1 |Ax|2 = (Ax)T(Ax) = (xTAT)(Ax) = xT(ATA)x Expand x in terms of eigenvectors of ATA: x = m1e1+ m2e2+… xT(ATA)x = l1m12+l2m22+… |x|2 = m12+m22+…

Camera Calibration – Example 1 To minimize l1m12+l2m22+… subject to m12+m22+… = 1 set mmin= 1 and all other mi=0 Thus, least squares solution is eigenvector corresponding to minimum (non-zero) eigenvalue of ATA

Camera Calibration – Example 2 Incorporating additional constraints into camera model No shear (u, v axes orthogonal) Square pixels etc. Doing minimization in image space All of these impose nonlinear constraints on camera parameters

Camera Calibration – Example 2 Option 1: nonlinear least squares Usually “gradient descent” techniques e.g. Levenberg-Marquardt Option 2: solve for general perspective model, find closest solution that satisfies constraints Use closed-form solution as initial guess for iterative minimization

Radial Distortion Radial distortion can not be represented by matrix (cu, cv) is image center, u*img= uimg– cu, v*img= vimg– cv, k is first-order radial distortion coefficient

Camera Calibration – Example 3 Incorporating radial distortion Option 1: Find distortion first (e.g., straight lines in calibration target) Warp image to eliminate distortion Run (simpler) perspective calibration Option 2: nonlinear least squares

Calibration Targets Full 3D (nonplanar) 2D (planar) Can calibrate with one image Difficult to construct 2D (planar) Can be made more accuracte Need multiple views Better constrained than full SFM problem

Calibration Targets Identification of features Manual Regular array, manually seeded Regular array, automatically seeded Color coding, patterns, etc. Subpixel estimation of locations Circle centers Checkerboard corners

Calibration Target w. Circles

3D Target w. Circles

Planar Checkerboard Target [Bouguet]

Coded Circles [Marschner et al.] Here’s a photo of the actual setup. You can see the camera on the left; we move this during the measurement session as we capture images. The light source is on the right; it’s a studio-grade electronic flash in a Fome-Cor box we built to resuce stray light. In the middle is the test object: in this case my daughter, who just turned 10 this week. You can also see a lot of little white dots, which I’ll explain next. [Marschner et al.]

Concentric Coded Circles [Gortler et al.]

Color Coded Circles [Culbertson]

Calibrating Projector Calibrate camera Project pattern onto a known object (usually plane) Can use time-coded structured light Form (uproj, vproj, x, y, z) tuples Use regular camera calibration code Typically lots of keystoning relative to cameras

Multi-Camera Geometry Epipolar geometry – relationship between observed positions of points in multiple cameras Assume: 2 cameras Known intrinsics and extrinsics

Epipolar Geometry P p1 p2 C1 C2

Epipolar Geometry P l2 p1 p2 C1 C2

Epipolar Geometry P Epipolar line l2 p1 p2 C1 C2 Epipoles

Epipolar Geometry Goal: derive equation for l2 Observation: P, C1, C2 determine a plane P l2 p1 p2 C1 C2

Epipolar Geometry Work in coordinate frame of C1 Normal of plane is T  Rp2, where T is relative translation, R is relative rotation P l2 p1 p2 C1 C2

Epipolar Geometry p1 is perpendicular to this normal: p1  (T  Rp2) = 0 P l2 p1 p2 C1 C2

Epipolar Geometry Write cross product as matrix multiplication P C1 C2

Epipolar Geometry p1  T* R p2 = 0  p1T E p2 = 0 E is the essential matrix P l2 p1 p2 C1 C2

Essential Matrix E depends only on camera geometry Given E, can derive equation for line l2 P l2 p1 p2 C1 C2

Fundamental Matrix Can define fundamental matrix F analogously, operating on pixel coordinates instead of camera coordinates u1T F u2 = 0 Advantage: can sometimes estimate F without knowing camera calibration