Algorithms & Applications in Computer Vision

Slides:



Advertisements
Similar presentations
Structure from motion.
Advertisements

The fundamental matrix F
Lecture 11: Two-view geometry
3D reconstruction.
Stereo and Projective Structure from Motion
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Two-View Geometry CS Sastry and Yang
N-view factorization and bundle adjustment CMPUT 613.
Self-calibration.
Structure from motion Class 9 Read Chapter 5. 3D photography course schedule (tentative) LectureExercise Sept 26Introduction- Oct. 3Geometry & Camera.
Two-view geometry.
Lecture 11: Structure from Motion
Camera calibration and epipolar geometry
Structure from motion.
Projective structure from motion
Computer Vision Projective structure from motion Marc Pollefeys COMP 256 Some slides and illustrations from J. Ponce, …
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Multiple-view Reconstruction from Points and Lines
3D reconstruction of cameras and structure x i = PX i x’ i = P’X i.
Many slides and illustrations from J. Ponce
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
Stereo and Structure from Motion
© 2003 by Davi GeigerComputer Vision October 2003 L1.1 Structure-from-EgoMotion (based on notes from David Jacobs, CS-Maryland) Determining the 3-D structure.
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
Computer Vision Structure from motion Marc Pollefeys COMP 256 Some slides and illustrations from J. Ponce, A. Zisserman, R. Hartley, Luc Van Gool, …
Affine structure from motion
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
Lecture 12: Structure from motion CS6670: Computer Vision Noah Snavely.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
Automatic Camera Calibration
Computer vision: models, learning and inference
Euclidean cameras and strong (Euclidean) calibration Intrinsic and extrinsic parameters Linear least-squares methods Linear calibration Degenerate point.
Multi-view geometry.
Homogeneous Coordinates (Projective Space) Let be a point in Euclidean space Change to homogeneous coordinates: Defined up to scale: Can go back to non-homogeneous.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
Structure from Motion Computer Vision CS 143, Brown James Hays 11/18/11 Many slides adapted from Derek Hoiem, Lana Lazebnik, Silvio Saverese, Steve Seitz,
Structure from Motion Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/06/12 Many slides adapted from Lana Lazebnik, Silvio Saverese,
Structure from Motion, Feature Tracking, and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 04/15/10 Many slides adapted.
CSCE 643 Computer Vision: Structure from Motion
Computer Vision – Lecture 17
Perceptual and Sensory Augmented Computing Computer Vision WS 08/09 Computer Vision – Lecture 17 Structure-from-Motion Bastian Leibe RWTH Aachen.
Affine Structure from Motion
Single-view geometry Odilon Redon, Cyclops, 1914.
EECS 274 Computer Vision Affine Structure from Motion.
776 Computer Vision Jan-Michael Frahm & Enrique Dunn Spring 2013.
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
Structure from Motion ECE 847: Digital Image Processing
Uncalibrated reconstruction Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration.
Structure from Motion Paul Heckbert, Nov , Image-Based Modeling and Rendering.
Structure from motion Multi-view geometry Affine structure from motion Projective structure from motion Planches : –
EECS 274 Computer Vision Projective Structure from Motion.
Lecture 22: Structure from motion CS6670: Computer Vision Noah Snavely.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
CSE 185 Introduction to Computer Vision Stereo 2.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
Reconstruction of a Scene with Multiple Linearly Moving Objects Mei Han and Takeo Kanade CISC 849.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
Motion Segmentation with Missing Data using PowerFactorization & GPCA
Two-view geometry Computer Vision Spring 2018, Lecture 10
Epipolar geometry.
Structure from motion Input: Output: (Tomasi and Kanade)
Uncalibrated Geometry & Stratification
Noah Snavely.
Multi-view geometry.
Structure from motion.
Structure from motion Input: Output: (Tomasi and Kanade)
Presentation transcript:

Algorithms & Applications in Computer Vision Lihi Zelnik-Manor Lecture 11: Structure from Motion

Structure from motion Stereo Multi-view

Two-frame SFM Calibrated case: Non-Calibrated case: Grauman 3

Recovering translation E is singular: SVD: Is the singular vector associated with the smallest eigenvalue We can only find it up to a scale and sign!!

Recovering T and R SVD: One solution: There are two opposite directions which are possible for  t  and two different rotations which are compatible with them. (change t to –t, and W into W-1)

Recovering T and R Given a pair of corresponding image coordinates, three of the solutions will always produce a 3D point which lies behind at least one of the two cameras and therefore cannot be seen. Only one of the four solutions will consistently produce 3D points which are in front of both cameras. This must then be the correct solution. Still, however, it has an undetermined positive scaling related to the translation component.

Structure from motion Lazebnik

Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding points in 3D? Correspondence (stereo matching): Given a point in just one image, how does it constrain the position of the corresponding point in another image? Camera geometry (motion): Given a set of corresponding points in two or more images, what are the camera matrices for these views? Lazebnik

Structure from motion Given: m images of n fixed 3D points xij = Pi Xj , i = 1, … , m, j = 1, … , n Problem: estimate m projection matrices Pi and n 3D points Xj from the mn correspondences xij x1j x2j x3j Xj P1 P2 P3 n points m images Lazebnik

Structure from motion ambiguity If we scale the entire scene by some factor k and, at the same time, scale the camera matrices by the factor of 1/k, the projections of the scene points in the image remain exactly the same: It is impossible to recover the absolute scale of the scene! Lazebnik

Structure from motion ambiguity If we scale the entire scene by some factor k and, at the same time, scale the camera matrices by the factor of 1/k, the projections of the scene points in the image remain exactly the same More generally: if we transform the scene using a transformation Q and apply the inverse transformation to the camera matrices, then the images do not change Lazebnik

Projective ambiguity Lazebnik

Projective ambiguity Lazebnik

Affine ambiguity Affine Lazebnik

Affine ambiguity Lazebnik

Similarity ambiguity Lazebnik

Similarity ambiguity Lazebnik

Hierarchy of 3D transformations Projective 15dof Preserves intersection and tangency Affine 12dof Preserves parallellism, volume ratios Similarity 7dof Preserves angles, ratios of length Euclidean 6dof Preserves angles, lengths With no constraints on the camera calibration matrix or on the scene, we get a projective reconstruction Need additional information to upgrade the reconstruction to affine, similarity, or Euclidean Lazebnik

Structure from motion Let’s start with affine cameras (the math is easier) center at infinity Lazebnik

Recall: Orthographic Projection Special case of perspective projection Distance from center of projection to image plane is infinite Projection matrix: Image World Lazebnik Slide by Steve Seitz

Affine cameras Orthographic Projection Parallel Projection Lazebnik

Projection of world origin Affine cameras A general affine camera combines the effects of an affine transformation of the 3D space, orthographic projection, and an affine transformation of the image: Affine projection is a linear mapping + translation in inhomogeneous coordinates x X a1 a2 Projection of world origin Lazebnik

Affine structure from motion Given: m images of n fixed 3D points: xij = Ai Xj + bi , i = 1,… , m, j = 1, … , n Problem: use the mn correspondences xij to estimate m projection matrices Ai and translation vectors bi, and n points Xj The reconstruction is defined up to an arbitrary affine transformation Q (12 degrees of freedom): We have 2mn knowns and 8m + 3n unknowns (minus 12 dof for affine ambiguity) Thus, we must have 2mn >= 8m + 3n – 12 For two views, we need four point correspondences Lazebnik

Getting rid of b Centering: subtract the centroid of the image points For simplicity, assume that the origin of the world coordinate system is at the centroid of the 3D points After centering, each normalized point xij is related to the 3D point Xi by Lazebnik

Affine structure from motion Let’s create a 2m × n data (measurement) matrix: cameras (2 m) points (n) C. Tomasi and T. Kanade. Shape and motion from image streams under orthography: A factorization method. IJCV, 9(2):137-154, November 1992.

Affine structure from motion Let’s create a 2m × n data (measurement) matrix: points (3 × n) cameras (2 m × 3) The measurement matrix D = MS must have rank 3! C. Tomasi and T. Kanade. Shape and motion from image streams under orthography: A factorization method. IJCV, 9(2):137-154, November 1992.

Factorizing the measurement matrix Lazebnik Source: M. Hebert

Factorizing the measurement matrix Singular value decomposition of D: Lazebnik Source: M. Hebert

Factorizing the measurement matrix Singular value decomposition of D: Lazebnik Source: M. Hebert

Factorizing the measurement matrix Obtaining a factorization from SVD: Lazebnik Source: M. Hebert

Factorizing the measurement matrix Obtaining a factorization from SVD: This decomposition minimizes |D-MS|2 Lazebnik Source: M. Hebert

Affine ambiguity The decomposition is not unique. We get the same D by using any 3×3 matrix C and applying the transformations M → MC, S →C-1S That is because we have only an affine transformation and we have not enforced any Euclidean constraints (like forcing the image axes to be perpendicular, for example) Lazebnik Source: M. Hebert

Eliminating the affine ambiguity Orthographic: image axes are perpendicular and scale is 1 This translates into 3m equations in L = CCT : Ai L AiT = I, i = 1, …, m Solve for L Recover C from L by Cholesky decomposition: L = CCT Update M and S: M = MC, S = C-1S a1 · a2 = 0 x |a1|2 = |a2|2 = 1 a2 X a1 Lazebnik Source: M. Hebert

Algorithm summary Given: m images and n features xij For each image i, center the feature coordinates Construct a 2m × n measurement matrix D: Column j contains the projection of point j in all views Row i contains one coordinate of the projections of all the n points in image i Factorize D: Compute SVD: D = U W VT Create U3 by taking the first 3 columns of U Create V3 by taking the first 3 columns of V Create W3 by taking the upper left 3 × 3 block of W Create the motion and shape matrices: M = U3W3½ and S = W3½ V3T (or M = U3 and S = W3V3T) Eliminate affine ambiguity Lazebnik Source: M. Hebert

Reconstruction results C. Tomasi and T. Kanade. Shape and motion from image streams under orthography: A factorization method. IJCV, 9(2):137-154, November 1992.

Dealing with missing data So far, we have assumed that all points are visible in all views In reality, the measurement matrix typically looks something like this: cameras points Lazebnik

Dealing with missing data Possible solution: decompose matrix into dense sub-blocks, factorize each sub-block, and fuse the results Finding dense maximal sub-blocks of the matrix is NP-complete (equivalent to finding maximal cliques in a graph) Incremental bilinear refinement Perform factorization on a dense sub-block (2) Solve for a new 3D point visible by at least two known cameras (linear least squares) (3) Solve for a new camera that sees at least three known 3D points (linear least squares) F. Rothganger, S. Lazebnik, C. Schmid, and J. Ponce. Segmenting, Modeling, and Matching Video Clips Containing Multiple Moving Objects. PAMI 2007.

Further Factorization work Factorization with uncertainty Factorization for indep. moving objects Factorization for articulated objects Factorization for dynamic objects Perspective factorization Factorization with outliers and missing pts. (Irani & Anandan, IJCV’02) (Costeira and Kanade ‘94) (Yan and Pollefeys ‘05) (Bregler et al. 2000, Brand 2001) (Sturm & Triggs 1996, …) (Jacobs ‘97 (affine), Martinek & Pajdla‘01 Aanaes’02 (perspective)) Pollefeys

Multi-Body Factorization Single body (Tomasi & Kanade): Multi-body (Costeira & Kanade, Gear): Block diagonal The shape interaction matrix: is block diagonal!!

Multi-Body Factorization Single body (Tomasi & Kanade): Multi-body (Costeira & Kanade, Gear): Block diagonal Pros: Fast and simple Cons: Carefully tracked (sparse) feature points Only rigid objects

Multi-Body Factorization Pollefeys

Non-independent motions Linear independence Partial dependence Segmentation requires block diagonal structure Wrong segmentation

Shape interaction matrix Shape interaction matrix for articulated objects looses block diagonal structure Costeira and Kanade’s approach is not usable for articulated bodies (assumes independent motions) Pollefeys

Articulated motion subspaces Motion subspaces for articulated bodies intersect (Yan and Pollefeys, CVPR’05) (Tresadern and Reid, CVPR’05) Joint (1D intersection) (joint=origin) (rank=8-1) Hinge (2D intersection) (hinge=z-axis) (rank=8-2) Exploit rank constraint to obtain better estimate Also for non-rigid parts if (Yan & Pollefeys, 06?)

Results Toy truck Student Segmentation Segmentation Intersection Pollefeys

Structure from motion of deforming objects (Bregler et al ’00; Brand ‘01) Extend factorization approaches to deal with dynamic shapes http://movement.stanford.edu/automatic-nr-modeling/ Pollefeys

Representing dynamic shapes (fig. M.Brand) represent dynamic shape as varying linear combination of basis shapes Pollefeys

Projecting dynamic shapes (figs. M.Brand) Rewrite: Pollefeys

Dynamic image sequences One image: (figs. M.Brand) Multiple images Pollefeys

Dynamic SfM factorization? Problem: find J so that M has proper structure Pollefeys

Dynamic SfM factorization (Bregler et al ’00) Assumption: SVD preserves order and orientation of basis shape components Pollefeys

Results (Bregler et al ’00) Pollefeys

Projective structure from motion Given: m images of n fixed 3D points zij xij = Pi Xj , i = 1,… , m, j = 1, … , n Problem: estimate m projection matrices Pi and n 3D points Xj from the mn correspondences xij x1j x2j x3j Xj P1 P2 P3 Lazebnik

Projective structure from motion Given: m images of n fixed 3D points zij xij = Pi Xj , i = 1,… , m, j = 1, … , n Problem: estimate m projection matrices Pi and n 3D points Xj from the mn correspondences xij With no calibration info, cameras and points can only be recovered up to a 4x4 projective transformation Q: X → QX, P → PQ-1 We can solve for structure and motion when 2mn >= 11m +3n – 15 For two cameras, at least 7 points are needed Lazebnik

Projective SFM: Two-camera case Compute fundamental matrix F between the two views First camera matrix: [I|0] Second camera matrix: [A|b] Then b: epipole (FTb = 0), A = –[b×]F Lazebnik F&P sec. 13.3.1

Projective factorization points (4 × n) cameras (3 m × 4) D = MS has rank 4 If we knew the depths z, we could factorize D to estimate M and S If we knew M and S, we could solve for z Solution: iterative approach (alternate between above two steps) Lazebnik

Sequential structure from motion Initialize motion from two images using fundamental matrix Initialize structure For each additional view: Determine projection matrix of new camera using all the known 3D points that are visible in its image – calibration points cameras Lazebnik

Sequential structure from motion Initialize motion from two images using fundamental matrix Initialize structure For each additional view: Determine projection matrix of new camera using all the known 3D points that are visible in its image – calibration Refine and extend structure: compute new 3D points, re-optimize existing points that are also seen by this camera – triangulation points cameras Lazebnik

Sequential structure from motion Initialize motion from two images using fundamental matrix Initialize structure For each additional view: Determine projection matrix of new camera using all the known 3D points that are visible in its image – calibration Refine and extend structure: compute new 3D points, re-optimize existing points that are also seen by this camera – triangulation Refine structure and motion: bundle adjustment points cameras Lazebnik

Bundle adjustment Non-linear method for refining structure and motion Minimizing reprojection error Xj P1Xj x3j x1j P3Xj P2Xj x2j P1 P3 P2 Lazebnik

Summary: Structure from motion Ambiguity Affine structure from motion: factorization Dealing with missing data Projective structure from motion: two views Projective structure from motion: iterative factorization Bundle adjustment Self-calibration Lazebnik

Photo Tourism: Exploring Photo Collections in 3D (Included presentation…available from http://phototour.cs.washington.edu/) Photo Tourism: Exploring Photo Collections in 3D Noah Snavely Steven M. Seitz University of Washington Richard Szeliski Microsoft Research © 2006 Noah Snavely

http://grail.cs.washington.edu/rome/

“Rome in a day”: Coliseum video http://grail.cs.washington.edu/rome/

“Rome in a day”: Trevi video http://grail.cs.washington.edu/rome/

“Rome in a day”: St. Peters video http://grail.cs.washington.edu/rome/

Moving to Multiple sequences What are the effects of dependencies between and ?

Single Vs. Multiple Sequences Columns Rows Dependence between: Motions Shapes Single sequence Wrong segmentation No effect Information for spatial matching (across sequences) Information for temporal synchronization Multiple sequences

Spatial Matching … but … only if the columns are ordered correctly shapes Dependence between: rows Point n Point n+1 Point n+2 rank … but … only if the columns are ordered correctly permutation Minimal rank Point n+2 Point n Point n+1 Spatial matching

Spatial Matching – When? Requires shape dependence only Same points, different motions Sequences of different length

Temporal Synchronization motions Dependence between: columns … but … only if the rows are ordered correctly

Temporal Synchronization motions Dependence between: columns Minimal rank Temporal synchronization

Temporal Synchronization – When? Requires motion dependence only Different points in each sequence Partial linear dependence suffices Partial spatial overlap A small number of points suffices Seq1 Seq2

Temporal Synchronization Result

Slide Credits Trevor Darrell Svetlana Lazebnik Marc Pollefeys Noah Snaveley & co-authors