Advanced Computer Vision Chapter 11 Stereo Correspondence Presented by: 蘇唯誠 0921679513 指導教授 : 傅楸善 博士.

Slides:



Advertisements
Similar presentations
1Ellen L. Walker Stereo Vision Why? Two images provide information to extract (some) 3D information We have good biological models (our own vision system)
Advertisements

Lecture 11: Two-view geometry
CSE473/573 – Stereo and Multiple View Geometry
Stereo Vision Reading: Chapter 11
Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Lecture 8: Stereo.
Stereo.
Last Time Pinhole camera model, projection
Stanford CS223B Computer Vision, Winter 2005 Lecture 6: Stereo 2 Sebastian Thrun, Stanford Rick Szeliski, Microsoft Hendrik Dahlkamp and Dan Morris, Stanford.
Computer Vision : CISC 4/689 Adaptation from: Prof. James M. Rehg, G.Tech.
Stereo Matching Vision for Graphics CSE 590SS, Winter 2001 Richard Szeliski.
Stereo & Iterative Graph-Cuts Alex Rav-Acha Vision Course Hebrew University.
Multi-view stereo Many slides adapted from S. Seitz.
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
Stereopsis Mark Twain at Pool Table", no date, UCR Museum of Photography.
The plan for today Camera matrix
CS 223b 1 More on stereo and correspondence. CS 223b 2 =?f g Mostpopular For each window, match to closest window on epipolar line in other image. (slides.
Stereo Computation using Iterative Graph-Cuts
Lecture 11: Stereo and optical flow CS6670: Computer Vision Noah Snavely.
Lec 21: Fundamental Matrix
CSE473/573 – Stereo Correspondence
Announcements PS3 Due Thursday PS4 Available today, due 4/17. Quiz 2 4/24.
Stereo Sebastian Thrun, Gary Bradski, Daniel Russakoff Stanford CS223B Computer Vision (with slides by James Rehg and.
Visualization- Determining Depth From Stereo Saurav Basu BITS Pilani 2002.
Stereo matching “Stereo matching” is the correspondence problem –For a point in Image #1, where is the corresponding point in Image #2? C1C1 C2C2 ? ? C1C1.
Stereo matching Class 10 Read Chapter 7 Tsukuba dataset.
Stereo vision A brief introduction Máté István MSc Informatics.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
Automatic Camera Calibration
Epipolar Geometry and Stereo Vision Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/05/15 Many slides adapted from Lana Lazebnik,
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
Brief Introduction to Geometry and Vision
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
Epipolar geometry Epipolar Plane Baseline Epipoles Epipolar Lines
Stereo Vision Reading: Chapter 11 Stereo matching computes depth from two or more images Subproblems: –Calibrating camera positions. –Finding all corresponding.
Stereo Many slides adapted from Steve Seitz.
Stereo Many slides adapted from Steve Seitz. Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image image 1image.
Computer Vision, Robert Pless
Advanced Computer Vision Chapter 11 Stereo Correspondence Presented by: 傅楸善 & 李知駿 指導教授 : 傅楸善 博士.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Computer Vision Stereo Vision. Bahadir K. Gunturk2 Pinhole Camera.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Bahadir K. Gunturk1 Phase Correlation Bahadir K. Gunturk2 Phase Correlation Take cross correlation Take inverse Fourier transform  Location of the impulse.
Lecture 16: Stereo CS4670 / 5670: Computer Vision Noah Snavely Single image stereogram, by Niklas EenNiklas Een.
stereo Outline : Remind class of 3d geometry Introduction
Feature Matching. Feature Space Outlier Rejection.
Solving for Stereo Correspondence Many slides drawn from Lana Lazebnik, UIUC.
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
John Morris Stereo Vision (continued) Iolanthe returns to the Waitemata Harbour.
Geometry Reconstruction March 22, Fundamental Matrix An important problem: Determine the epipolar geometry. That is, the correspondence between.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
CSE 185 Introduction to Computer Vision Stereo 2.
Stereo CS4670 / 5670: Computer Vision Noah Snavely Single image stereogram, by Niklas EenNiklas Een.
CS 6501: 3D Reconstruction and Understanding Stereo Cameras
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
Summary of “Efficient Deep Learning for Stereo Matching”
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
STEREOPSIS The Stereopsis Problem: Fusion and Reconstruction
Advanced Computer Vision Chapter 11 Stereo Correspondence
Geometry 3: Stereo Reconstruction
EECS 274 Computer Vision Stereopsis.
What have we learned so far?
Course 6 Stereo.
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Stereo vision Many slides adapted from Steve Seitz.
Presentation transcript:

Advanced Computer Vision Chapter 11 Stereo Correspondence Presented by: 蘇唯誠 指導教授 : 傅楸善 博士

Introduction Stereo matching is the process of taking two or more images and estimating a 3D model of the scene by finding matching pixels in the images and converting their 2D positions into 3D depths. In this chapter, we address the question of how to build a more complete 3D model.

(a) (b) (c) (d) (e) (f)

11.1 Epipolar Geometry Given a pixel in one image, how can we compute its correspondence in the other image? We exploit this information to reduce the number of potential correspondences, and hence both speed up the matching and increase its reliability.

(Faugeras and Luong 2001; Hartley and Zisserman 2004).

Rectification We can use the epipolar line corresponding to a pixel in one image to constrain the search for corresponding pixels in the other image. One way to do this is to use a general correspondence algorithm, such as optical flow. A more efficient algorithm can be obtained by first rectifying (i.e, warping).

(a) (b) (c) (d) (a) Original image pair overlaid with several epipolar lines; (b) images transformed so that epipolar lines are parallel; (c) images rectified so that epipolar lines are horizontal and in vertial correspondence; (d) final rectification that minimizes horizontal distortions. (Loop and Zhang 1999; Faugeras and Luong 2001; Hartley and Zisserman 2004).

Rectification

Rectified images have the following two properties: 1.All epipolar lines are parallel to the u- coordinate axis 2.Corresponding points have identical v- coordinates

Rectification

(a) (b) (c) (d) (a) Original image pair overlaid with several epipolar lines; (b) images transformed so that epipolar lines are parallel; (c) images rectified so that epipolar lines are horizontal and in vertial correspondence; (d) final rectification that minimizes horizontal distortions. (Loop and Zhang 1999; Faugeras and Luong 2001; Hartley and Zisserman 2004).

Rectification The resulting standard rectified geometry is employed in a lot of stereo camera setups and stereo algorithms, and leads to a very simple inverse relationship between 3D depths Z and disparities d, where f is the focal length (measured in pixels), B is the baseline, and

Rectification The task of extracting depth from a set of images then becomes one of estimating the disparity map d(x, y). After rectification, we can easily compare the similarity of pixels at corresponding locations (x, y) and (x 0, y 0 ) = (x + d, y) and store them in a disparity space image C(x, y, d) for further processing. DSI: disparity space image

Bobick and Intille (1999).

11.2 Sparse correspondence Early stereo matching algorithms were feature-based, i.e., they first extracted a set of potentially matchable image locations, using either interest operators or edge detectors, and then searched for corresponding locations in other images using a patch-based metric. (Hannah 1974; Marr and Poggio 1979; Mayhew and Frisby 1980; Baker and Binford 1981; Arnold 1983; Grimson 1985; Ohta and Kanade 1985; Bolles, Baker, and Marimont 1987; Matthies, Kanade, and Szeliski 1989; Hsieh, McKeown, and Perlant 1992; Bolles, Baker, and Hannah 1993).

D curves and profiles Another example of sparse correspondence is the matching of profile curves. Let us assume that the camera is moving smoothly enough that the local epipolar geometry varies slowly.

(a) (b) (c) (d) (e) (f) (g) (a) circular arc fitting in the epipolar plane; (b) synthetic example of an ellipsoid with a truncated side and elliptic surface markings; (c) partially reconstructed surface mesh seen from an oblique and top-down view; (d) real-world image sequence of a soda can on a turntable; (e) extracted edges; (f) partially reconstructed profile curves; (g) partially reconstructed surface mesh.

D curves and profiles

11.3 Dense correspondence While sparse matching algorithms are still occasionally used, most stereo matching algorithms today focus on dense correspondence, since this is required for applications such as image-based rendering or modeling. This problem is more challenging than sparse correspondence.

11.3 Dense correspondence (cont’) It is based on the observation that stereo algorithms generally perform some subset of the following four steps: – 1. matching cost computation; – 2. cost (support) aggregation; – 3. disparity computation and optimization; and – 4. disparity refinement.

11.3 Dense correspondence (cont’) For example, the traditional sum-of-squared differences (SSD) algorithm can be described as: – 1. The matching cost is the squared difference of intensity values at a given disparity. – 2. Aggregation is done by summing the matching cost over square windows with constant disparity. – 3. Disparities are computed by selecting the minimal (winning) aggregated value at each pixel.

11.3 Dense correspondence (cont’) Global algorithms, on the other hand, make explicit smoothness assumptions and then solve a global optimization problem (Section 11.5). Such algorithms typically do not perform an aggregation step, but rather seek a disparity assignment (step 3) that minimizes a global cost function that consists of data (step 1) terms and smoothness terms.

11.4 Local methods Local and window-based methods aggregate the matching cost by summing or averaging over a support region in the DSI C(x, y, d). DSI: Disparity Space Image Aggregation has been implemented using square windows or Gaussian convolution (traditional).

11.5 Global optimization Global stereo matching methods perform some optimization or iteration steps after the disparity computation phase and often skip the aggregation step altogether. Many global methods are formulated in an energy- minimization framework.

11.5 Global optimization (cont’) The objective is to find a solution d that minimizes a global energy, The data term, E d (d), measures how well the disparity function d agrees with the input image pair. where C is the (initial or aggregated) matching cost DSI.

11.5 Global optimization (cont’) The smoothness term is often restricted to measuring only the differences between neighboring pixels’ disparities, where ᵨ is some monotonically increasing function of disparity difference.

11.6 Multi-view stereo While matching pairs of images is a useful way of obtaining depth information, matching more images can lead to even better results.

11.6 Multi-view stereo

Automated Stereo Perception Following introduces the subproblem of one- dimensional stereo matching. Consider a pair of corresponding epipolar lines from a stereo image pair.

Automated Stereo Perception Rays from the left and right cameras intersect to from a grid or lattice. This lattice is bounded by a region obtained by projecting rays through the image boundaries. We will refer to this region as the stereo zone, for only objects within this zone can be seen in stereo.

Automated Stereo Perception

Each lattice point corresponds to a potential match between a feature in the left and a feature in the right. If such a match were correct, then the object must have been at the point space represented by that lattice point.

Automated Stereo Perception We have six types of profile intervals: 1.The surface and both edges are visible to both cameras. 2.The surface and its left edge are visible to both cameras, but its right edge is occluded. 3.The surface is visible only to the left camera. 4.The surface is visible only to the right camera. 5.The surface and its right edge are visible to both cameras, but its left edge is occluded. 6.The surface is visible to both cameras, but its left and right edges are occluded.

Automated Stereo Perception

Summary: – This section introduced some basic stereo matching. – We describe the geometry and develop a notation for a stereo system.

END