Stereo vision A brief introduction Máté István MSc Informatics.

Slides:



Advertisements
Similar presentations
Epipolar Geometry.
Advertisements

Multiview Reconstruction. Why More Than 2 Views? BaselineBaseline – Too short – low accuracy – Too long – matching becomes hard.
www-video.eecs.berkeley.edu/research
CS 376b Introduction to Computer Vision 04 / 21 / 2008 Instructor: Michael Eckmann.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
3D Computer Vision and Video Computing 3D Vision Topic 3 of Part II Stereo Vision CSc I6716 Spring 2011 Zhigang Zhu, City College of New York
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Last Time Pinhole camera model, projection
Computer Vision : CISC 4/689
Multi-view stereo Many slides adapted from S. Seitz.
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
Stereopsis Mark Twain at Pool Table", no date, UCR Museum of Photography.
3D Computer Vision and Video Computing 3D Vision Topic 4 of Part II Stereo Vision CSc I6716 Spring 2008 Zhigang Zhu, City College of New York
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
3D Computer Vision and Video Computing 3D Vision Lecture 15 Stereo Vision (II) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
CSE473/573 – Stereo Correspondence
December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding.
Announcements PS3 Due Thursday PS4 Available today, due 4/17. Quiz 2 4/24.
Visualization- Determining Depth From Stereo Saurav Basu BITS Pilani 2002.
COMP322/S2000/L271 Stereo Imaging Ref.V.S.Nalwa, A Guided Tour of Computer Vision, Addison Wesley, (ISBN ) Slides are adapted from CS641.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2006 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #15.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
What Does the Scene Look Like From a Scene Point? Donald Tanguay August 7, 2002 M. Irani, T. Hassner, and P. Anandan ECCV 2002.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
John Morris These slides were adapted from a set of lectures written by Mircea Nicolescu, University of Nevada at Reno Stereo Vision Iolanthe in the Bay.
Shape from Stereo  Disparity between two images  Photogrammetry  Finding Corresponding Points Correlation based methods Feature based methods.
CS654: Digital Image Analysis Lecture 8: Stereo Imaging.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
CSCE 643 Computer Vision: Structure from Motion
Plenoptic Modeling: An Image-Based Rendering System Leonard McMillan & Gary Bishop SIGGRAPH 1995 presented by Dave Edwards 10/12/2000.
Lecture 03 15/11/2011 Shai Avidan הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Computer Vision Stereo Vision. Bahadir K. Gunturk2 Pinhole Camera.
Stereo Vision John Morris These slides were adapted from a set of lectures written by Mircea Nicolescu, University of Nevada at Reno Vision Research in.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
Bahadir K. Gunturk1 Phase Correlation Bahadir K. Gunturk2 Phase Correlation Take cross correlation Take inverse Fourier transform  Location of the impulse.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
Geometric Transformations
stereo Outline : Remind class of 3d geometry Introduction
(c) 2000, 2001 SNU CSE Biointelligence Lab Finding Region Another method for processing image  to find “regions” Finding regions  Finding outlines.
Stereo Vision John Morris Vision Research in CITR
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
October 19, STEREO IMAGING (continued). October 19, RIGHT IMAGE PLANE LEFT IMAGE PLANE RIGHT FOCAL POINT LEFT FOCAL POINT BASELINE d FOCAL.
3D Reconstruction Using Image Sequence
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
John Morris Stereo Vision (continued) Iolanthe returns to the Waitemata Harbour.
Geometry Reconstruction March 22, Fundamental Matrix An important problem: Determine the epipolar geometry. That is, the correspondence between.
Advanced Computer Vision Chapter 11 Stereo Correspondence Presented by: 蘇唯誠 指導教授 : 傅楸善 博士.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
John Morris These slides were adapted from a set of lectures written by Mircea Nicolescu, University of Nevada at Reno Stereo Vision Iolanthe in the Bay.
Media Processor Lab. Media Processor Lab. Trellis-based Parallel Stereo Matching Media Processor Lab. Sejong univ.
Lec 26: Fundamental Matrix CS4670 / 5670: Computer Vision Kavita Bala.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
CSE 185 Introduction to Computer Vision Stereo 2.
제 5 장 스테레오.
Computational Vision CSCI 363, Fall 2016 Lecture 15 Stereopsis
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Common Classification Tasks
Topic 7 of Part 2 Stereo Vision (II)
Reconstruction.
Computer Vision Stereo Vision.
Course 6 Stereo.
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Presentation transcript:

Stereo vision A brief introduction Máté István MSc Informatics

What the stereo vision aims Retrieving 3D information, and structure of an object with two, or one moving camera. In this project we use one moving camera. A line and a plane, not including it, intersect in just one point. Lines of sight are easy to compute, and so its easy to tell where any image point projects on to any known plane. If two images from different viewpoints can be placed in correspondence, the intersection of the lines of sight from two matching image points determines a point in 3D space.

Stereo vision and triangulation One of the first ideas that occurs to one who wants to do three-dimensional sensing is the biologically motivated one of stereo vision. Two cameras, or one from two positions, can give relative depth, or absolute three-dimensional location. There has been considerable effort in this direction [Moravec 1977, Quam and Hannah 1974, Binford 1971, Turner 1974, Shapira 1974]

The technique 1. Take two images separated by a baseline 2. Identify points between the two images 3. Use the inverse perspective transform, or simple triangulation to derive the two lines on witch the world points lie. 4. Intersect the lines The resulting point is in three-dimensional world coordinates. The hardest part of this is method is step 2, that of identifying corresponding points in the two images.

Stereo vision terminology Fixation point: the intersection of optical axis Baseline: the distance between the centers of the projection Epipolar plane: the plane passing through the conters of projection and the point in the scene Epipolar line: the intersection of the epipolar plane with the image plane Conjugate pair: any point in the scene that is visible in both cameras will be projected to a pair of image points in the two images

Disparity: the distance between corresponding points when the two images are superimposed Disparity map: the disparities of all points from the disparity map (can be displayed as an image)

Triangulation- the principle underlying stereo vision The 3D location of any visible object point in space is restricted to the straight line that passes trough the center of projection and projection of the object point Binocular stereo vision determines the position of a point in space by finding the intersection of the two lines passing through the center of projection an the projection of the point in each image

Two main problems of stereo vision I. The correspondence problem II. The reconstruction problem

Finding pairs of matched points such, that each point in the pair is the projection of the same 3D point Triangulation depends crucially on the solution of the correspondence problem. Ambiguous correspondence between points in the two images may lead to several different consistent interpretation of the scene I. The correspondence problem

Efficient correlation is of technological concern, but even if it were free and instantaneous, it would still be inadequate. The basic problems with correlation in stereo imaging relate to the fact that objects can look significantly different from different viewpoints It is possible for the two stereo views to be sufficiently different that corresponding areas may not be matched correctly Worse, in scenes with much obstruction, very important features of the scene may be present in only one view.

This problem is alleviated by decreasing the baseline, but the accuracy of depth determination suffers. One solution is to identify world features, not image appearance, in the two views, and match those (the nose of a person, the corner of a cube)

Why is the correspondence problem difficult? Some points in each image will have no corresponding point in the other image, bacause: The cameras may have different fields of view Due to occlusion A stereo system must be able to determine the image parts that should not be matched.

In the above picture, the part with green and red are the parts that show the different viewpoint of the cameras The task is to find points, that can be seen for both cameras Occlusion is both visible at the right edge of the box

Methods for establishing correspondence There are two issues to be considered: How to select candidate matches ? How to determine the goodness of a match? A possible class of algorithm Correlation based attempt to establish correspondence by matching image intensities

Correlation-based methods Match image sub-windows between the two images using image correlation Scene points must have the same intensity in each image (strictly accurate for perfectly matte surfaces only)

The algorithm Two images I L and I R are given In one of the images (I L ) consider a sub-window W, in the other a point P=(P x, P y ) The search region in the right image R(p L ) associated with a pixel p L in the left image For each pixel p L = (i, j) in the left image: For a displacement d=(dx,dy) in R ( p L ) find C(d) – a “norm” (Euclidian, Minkowski), correlation between the pixel pairs in images

Example: I choose the absolute difference between RGB pixel values: C(d) = SumAbs( PV (R,G,B) ( W ij )- PV (R,G,B) ( I R ( P x + i + dx ), I R ( P y + j + dy ))) This expresses that we count the difference The disparity of p L is the vector d’=(dx’,dy’) that minimizes C(d) over R(pr) d’ = arg min[C(d)] Improvements: I used edge-define in each picture, to produce more accurate results, given that I work with colored pictures in RGB space, so the algorithm is more like a feature-based algorithm without rotation and stretching

The pictures after edge-finding This way the matching is more accurate given that, mainly only edges remained

Problems and “how-to”-s This algorithm works well in a case of randomly given W sub-window, and a point P, that is that is at a chosen distance d=(Apx,Bpx) The question is how to determine the starting d, and the initial W sub-window, knowing that it can be a sub-window not present in the other image (due to the different camera viewpoint)

How to determine a threshold, to speed up computation How to determine the sub-window size? Too large sub-window becomes inaccurate, due rotation in the images, too small becomes inaccurate due lack of information To answer these questions intensive data observation and behavior is needed

The result for an arbitrary W sub-window and a P point. The result is quite good, but the image rotation can be seen

The reconstruction problem Given the corresponding points, we can compute the disparity map The disparity can be converted to a 3D map of the scene

Incorrect matching can give bad results

Recovering depth (reconstruction) Consider recovering the position of P from its projections p r, and p l Usually, the two cameras are related by the following transformation Using Z r = Z l = Z and X r = X l – T, we have where d = x l – x r is the disparity ( the difference in the position between the corresponding points in the two images )

The camera The camera is set on an old printer, that helps it moving in the same plane. This improves the stereo vision with single camera moving in a plane.

References [1] Computer vision - Dana Ballard, Christopher Brown [2] Stereo Camera - T. Kanade and M. Okutomi [3] Why use 3D data ? – Dave Marshall [4] Methods of 3D acquisition – Dave Marshall [5] The correspondence problem - T. Kanade and M. Okutomi