Last lecture Passive Stereo Spacetime Stereo.

Slides:



Advertisements
Similar presentations
Epipolar Geometry.
Advertisements

The fundamental matrix F
Lecture 11: Two-view geometry
3D reconstruction.
Structure from motion Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Zhengyou Zhang and Marc Pollefyes.
Structure from motion Digital Visual Effects, Spring 2005 Yung-Yu Chuang 2005/4/20 with slides by Richard Szeliski, Steve Seitz, Zhengyou Zhang and Marc.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Computer vision: models, learning and inference
Two-view geometry.
Lecture 8: Stereo.
Camera calibration and epipolar geometry
Structure from motion.
Robot Vision SS 2008 Matthias Rüther 1 ROBOT VISION Lesson 6: Shape from Stereo Matthias Rüther Slides partial courtesy of Marc Pollefeys Department of.
Epipolar Geometry Class 7 Read notes Feature tracking run iterative L-K warp & upsample Tracking Good features Multi-scale Transl. Affine.
Today Feature Tracking Structure from Motion on Monday (1/29)
Global Alignment and Structure from Motion
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
Lecture 21: Multiple-view geometry and structure from motion
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
Global Alignment and Structure from Motion Computer Vision CSE455, Winter 2008 Noah Snavely.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
Lecture 12: Structure from motion CS6670: Computer Vision Noah Snavely.
CS 558 C OMPUTER V ISION Lecture IX: Dimensionality Reduction.
Epipolar geometry Class 5. Geometric Computer Vision course schedule (tentative) LectureExercise Sept 16Introduction- Sept 23Geometry & Camera modelCamera.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
776 Computer Vision Jan-Michael Frahm, Enrique Dunn Spring 2013.
Automatic Camera Calibration
Computer vision: models, learning and inference
Multi-view geometry.
Epipolar geometry The fundamental matrix and the tensor
CSCE 643 Computer Vision: Structure from Motion
Multiview Geometry and Stereopsis. Inputs: two images of a scene (taken from 2 viewpoints). Output: Depth map. Inputs: multiple images of a scene. Output:
Stereo Course web page: vision.cis.udel.edu/~cv April 11, 2003  Lecture 21.
Robot Vision SS 2007 Matthias Rüther 1 ROBOT VISION Lesson 6a: Shape from Stereo, short summary Matthias Rüther Slides partial courtesy of Marc Pollefeys.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Announcements Project 3 due Thursday by 11:59pm Demos on Friday; signup on CMS Prelim to be distributed in class Friday, due Wednesday by the beginning.
Last lecture Passive Stereo Spacetime Stereo Multiple View Stereo.
Lecture 03 15/11/2011 Shai Avidan הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Geometry of Multiple Views
Affine Structure from Motion
Two-view geometry. Epipolar Plane – plane containing baseline (1D family) Epipoles = intersections of baseline with image planes = projections of the.
EECS 274 Computer Vision Affine Structure from Motion.
Structure from motion Digital Visual Effects, Spring 2008 Yung-Yu Chuang 2008/4/22 with slides by Richard Szeliski, Steve Seitz, Zhengyou Zhang and Marc.
776 Computer Vision Jan-Michael Frahm & Enrique Dunn Spring 2013.
Structure from Motion ECE 847: Digital Image Processing
Structure from Motion Paul Heckbert, Nov , Image-Based Modeling and Rendering.
Structure from motion Multi-view geometry Affine structure from motion Projective structure from motion Planches : –
Structure from motion Digital Visual Effects, Spring 2007 Yung-Yu Chuang 2007/4/24 with slides by Richard Szeliski, Steve Seitz, Zhengyou Zhang and Marc.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
Structure from motion Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Zhengyou Zhang and Marc Pollefyes.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
SFM under orthographic projection
Epipolar geometry.
3D Photography: Epipolar geometry
Structure from motion Input: Output: (Tomasi and Kanade)
Estimating 2-view relationships
Uncalibrated Geometry & Stratification
Reconstruction.
Two-view geometry.
Two-view geometry.
Digital Visual Effects, Spring 2006 Yung-Yu Chuang 2005/4/26
Structure from motion Input: Output: (Tomasi and Kanade)
Presentation transcript:

Last lecture Passive Stereo Spacetime Stereo

Today Structure from Motion: Given pixel correspondences, how to compute 3D structure and camera motion? Slides stolen from Prof Yungyu Chuang

Epipolar geometry & fundamental matrix

C,C’,x,x’ and X are coplanar The epipolar geometry epipolar geometry demo C,C’,x,x’ and X are coplanar

What if only C,C’,x are known? The epipolar geometry What if only C,C’,x are known?

All points on  project on l and l’ The epipolar geometry All points on  project on l and l’

The epipolar geometry Family of planes  and lines l and l’ intersect at e and e’

The epipolar geometry epipolar geometry demo epipolar pole = intersection of baseline with image plane = projection of projection center in other image epipolar geometry demo epipolar plane = plane containing baseline epipolar line = intersection of epipolar plane with image

The fundamental matrix F p’ p R C C’ T=C’-C Two reference frames are related via the extrinsic parameters The equation of the epipolar plane through X is

The fundamental matrix F essential matrix

The fundamental matrix F p’ p R C C’ T=C’-C

The fundamental matrix F Let M and M’ be the intrinsic matrices, then fundamental matrix

The fundamental matrix F p’ p R C C’ T=C’-C

The fundamental matrix F The fundamental matrix is the algebraic representation of epipolar geometry The fundamental matrix satisfies the condition that for any pair of corresponding points x↔x’ in the two images

The fundamental matrix F F is the unique 3x3 rank 2 matrix that satisfies x’TFx=0 for all x↔x’ Transpose: if F is fundamental matrix for (P,P’), then FT is fundamental matrix for (P’,P) Epipolar lines: l’=Fx & l=FTx’ Epipoles: on all epipolar lines, thus e’TFx=0, x e’TF=0, similarly Fe=0 F has 7 d.o.f. , i.e. 3x3-1(homogeneous)-1(rank2) F is a correlation, projective mapping from a point x to a line l’=Fx (not a proper correlation, i.e. not invertible) 右邊的epipole e為所有epipolar lines的交點, 亦即對於所有左邊image上的點x, xFe=0, it implies Fe=0. 亦即存在一個非零點使得Fv=0. 故e為F之null space, rank(null(F))=1 => rank(F)=2

The fundamental matrix F It can be used for Simplifies matching Allows to detect wrong matches

Estimation of F — 8-point algorithm The fundamental matrix F is defined by for any pair of matches x and x’ in two images. Let x=(u,v,1)T and x’=(u’,v’,1)T, each match gives a linear equation

8-point algorithm In reality, instead of solving , we seek f to minimize , least eigenvector of .

8-point algorithm It is achieved by SVD. Let , where , let To enforce that F is of rank 2, F is replaced by F’ that minimizes subject to . It is achieved by SVD. Let , where , let then is the solution.

8-point algorithm % Build the constraint matrix A = [x2(1,:)‘.*x1(1,:)' x2(1,:)'.*x1(2,:)' x2(1,:)' ... x2(2,:)'.*x1(1,:)' x2(2,:)'.*x1(2,:)' x2(2,:)' ... x1(1,:)' x1(2,:)' ones(npts,1) ]; [U,D,V] = svd(A); % Extract fundamental matrix from the column of V % corresponding to the smallest singular value. F = reshape(V(:,9),3,3)'; % Enforce rank2 constraint [U,D,V] = svd(F); F = U*diag([D(1,1) D(2,2) 0])*V';

8-point algorithm Pros: it is linear, easy to implement and fast Cons: susceptible to noise

Problem with 8-point algorithm ~10000 ~10000 ~100 ~10000 ~10000 ~100 ~100 ~100 1 Orders of magnitude difference between column of data matrix  least-squares yields poor results !

Normalized 8-point algorithm normalized least squares yields good results Transform image to ~[-1,1]x[-1,1] (0,500) (700,500) (-1,1) (1,1) (0,0) (0,0) (700,0) (-1,-1) (1,-1)

Normalized 8-point algorithm Transform input by , Call 8-point on to obtain

Normalized 8-point algorithm [x1, T1] = normalise2dpts(x1); [x2, T2] = normalise2dpts(x2); A = [x2(1,:)‘.*x1(1,:)' x2(1,:)'.*x1(2,:)' x2(1,:)' ... x2(2,:)'.*x1(1,:)' x2(2,:)'.*x1(2,:)' x2(2,:)' ... x1(1,:)' x1(2,:)' ones(npts,1) ]; [U,D,V] = svd(A); F = reshape(V(:,9),3,3)'; [U,D,V] = svd(F); F = U*diag([D(1,1) D(2,2) 0])*V'; % Denormalise F = T2'*F*T1;

Normalization function [newpts, T] = normalise2dpts(pts) c = mean(pts(1:2,:)')'; % Centroid newp(1,:) = pts(1,:)-c(1); % Shift origin to centroid. newp(2,:) = pts(2,:)-c(2); meandist = mean(sqrt(newp(1,:).^2 + newp(2,:).^2)); scale = sqrt(2)/meandist; T = [scale 0 -scale*c(1) 0 scale -scale*c(2) 0 0 1 ]; newpts = T*pts;

RANSAC compute F based on all inliers repeat select minimal sample (8 matches) compute solution(s) for F determine inliers until (#inliers,#samples)>95% or too many times compute F based on all inliers

Results (ground truth)

Results (8-point algorithm)

Results (normalized 8-point algorithm)

From F to R, T If we know camera parameters Hartley and Zisserman, Multiple View Geometry, 2nd edition, pp 259

CSE 576 (Spring 2005): Computer Vision Triangulation Problem: Given some points in correspondence across two or more images (taken from calibrated cameras), {(uj,vj)}, compute the 3D location X Richard Szeliski CSE 576 (Spring 2005): Computer Vision

CSE 576 (Spring 2005): Computer Vision Triangulation Method I: intersect viewing rays in 3D, minimize: X is the unknown 3D point Cj is the optical center of camera j Vj is the viewing ray for pixel (uj,vj) sj is unknown distance along Vj Advantage: geometrically intuitive X Vj Cj Richard Szeliski CSE 576 (Spring 2005): Computer Vision

CSE 576 (Spring 2005): Computer Vision Triangulation Method II: solve linear equations in X advantage: very simple Method III: non-linear minimization advantage: most accurate (image plane error) Richard Szeliski CSE 576 (Spring 2005): Computer Vision

Structure from motion

Structure from motion Unknown camera viewpoints structure for motion: automatic recovery of camera motion and scene structure from two or more images. It is a self calibration technique and called automatic camera tracking or matchmoving.

Applications For computer vision, multiple-view shape reconstruction, novel view synthesis and autonomous vehicle navigation. For film production, seamless insertion of CGI into live-action backgrounds

Structure from motion SFM pipeline 2D feature geometry optimization tracking 3D estimation optimization (bundle adjust) geometry fitting SFM pipeline

Structure from motion Step 1: Track Features Detect good features, Shi & Tomasi, SIFT Find correspondences between frames Lucas & Kanade-style motion estimation window-based correlation SIFT matching

Structure from Motion Step 2: Estimate Motion and Structure Simplified projection model, e.g., [Tomasi 92] 2 or 3 views at a time [Hartley 00]

Structure from Motion Step 3: Refine estimates “Bundle adjustment” in photogrammetry Other iterative methods

Structure from Motion Step 4: Recover surfaces (image-based triangulation, silhouettes, stereo…) Good mesh

Example : Photo Tourism

Factorization methods

Problem statement

SFM under orthographic projection matrix 2D image point 3D scene point image offset Trick Choose scene origin to be centroid of 3D points Choose image origins to be centroid of 2D points Allows us to drop the camera translation:

factorization (Tomasi & Kanade) projection of n features in one image: projection of n features in m images W measurement M motion S shape Key Observation: rank(W) <= 3

Factorization Factorization Technique known solve for Factorization Technique W is at most rank 3 (assuming no noise) We can use singular value decomposition to factor W: S’ differs from S by a linear transformation A: Solve for A by enforcing metric constraints on M

Metric constraints Orthographic Camera Rows of P are orthonormal: Enforcing “Metric” Constraints Compute A such that rows of M have these properties Trick (not in original Tomasi/Kanade paper, but in followup work) Constraints are linear in AAT : Solve for G first by writing equations for every Pi in M Then G = AAT by SVD

Results

Extensions to factorization methods Paraperspective [Poelman & Kanade, PAMI 97] Sequential Factorization [Morita & Kanade, PAMI 97] Factorization under perspective [Christy & Horaud, PAMI 96] [Sturm & Triggs, ECCV 96] Factorization with Uncertainty [Anandan & Irani, IJCV 2002]

Bundle adjustment

CSE 576 (Spring 2005): Computer Vision Structure from motion How many points do we need to match? 2 frames: (R,t): 5 dof + 3n point locations  4n point measurements  n  5 k frames: 6(k–1)-1 + 3n  2kn always want to use many more Richard Szeliski CSE 576 (Spring 2005): Computer Vision

CSE 576 (Spring 2005): Computer Vision Bundle Adjustment What makes this non-linear minimization hard? many more parameters: potentially slow poorer conditioning (high correlation) potentially lots of outliers Richard Szeliski CSE 576 (Spring 2005): Computer Vision

Lots of parameters: sparsity Only a few entries in Jacobian are non-zero Richard Szeliski CSE 576 (Spring 2005): Computer Vision

CSE 576 (Spring 2005): Computer Vision Robust error models Outlier rejection use robust penalty applied to each set of joint measurements for extremely bad data, use random sampling [RANSAC, Fischler & Bolles, CACM’81] Richard Szeliski CSE 576 (Spring 2005): Computer Vision

CSE 576 (Spring 2005): Computer Vision Correspondences Can refine feature matching after a structure and motion estimate has been produced decide which ones obey the epipolar geometry decide which ones are geometrically consistent (optional) iterate between correspondences and SfM estimates using MCMC [Dellaert et al., Machine Learning 2003] Richard Szeliski CSE 576 (Spring 2005): Computer Vision

Structure from motion: limitations Very difficult to reliably estimate metric structure and motion unless: large (x or y) rotation or large field of view and depth variation Camera calibration important for Euclidean reconstructions Need good feature tracker Lens distortion Richard Szeliski CSE 576 (Spring 2005): Computer Vision

Issues in SFM Track lifetime Nonlinear lens distortion Degeneracy and critical surfaces Prior knowledge and scene constraints Multiple motions

every 50th frame of a 800-frame sequence Track lifetime every 50th frame of a 800-frame sequence

lifetime of 3192 tracks from the previous sequence Track lifetime lifetime of 3192 tracks from the previous sequence

track length histogram Track lifetime track length histogram

Nonlinear lens distortion

Nonlinear lens distortion effect of lens distortion

Prior knowledge and scene constraints add a constraint that several lines are parallel

Prior knowledge and scene constraints add a constraint that it is a turntable sequence

Applications of Structure from Motion

Jurassic park

PhotoSynth http://labs.live.com/photosynth/