Presented by: Ali Agha March 02, 2009. Outline Sterevision overview Motivation & Contribution Structured light & method overview Related work Disparity.

Slides:



Advertisements
Similar presentations
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Advertisements

Structured Light principles Figure from M. Levoy, Stanford Computer Graphics Lab.
www-video.eecs.berkeley.edu/research
--- some recent progress Bo Fu University of Kentucky.
Depth from Structured Light II: Error Analysis
Gratuitous Picture US Naval Artillery Rangefinder from World War I (1918)!!
Stereo Many slides adapted from Steve Seitz. Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image Where does the.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
GrabCut Interactive Image (and Stereo) Segmentation Carsten Rother Vladimir Kolmogorov Andrew Blake Antonio Criminisi Geoffrey Cross [based on Siggraph.
Stereo.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Last Time Pinhole camera model, projection
CS6670: Computer Vision Noah Snavely Lecture 17: Stereo
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2005 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
Stereoscopic Light Stripe Scanning: Interference Rejection, Error Minimization and Calibration By: Geoffrey Taylor Lindsay Kleeman Presented by: Ali Agha.
Visibility Subspaces: Uncalibrated Photometric Stereo with Shadows Kalyan Sunkavalli, Harvard University Joint work with Todd Zickler and Hanspeter Pfister.
CS290 Spring 2000 Slide:1 Structured Light for Laparoscopic Surgery CS290 Computer Vision Jeremy Ackerman CS290 Computer Vision Jeremy Ackerman.
High Dynamic Range Imaging: Spatially Varying Pixel Exposures Shree K. Nayar, Tomoo Mitsunaga CPSC 643 Presentation # 2 Brien Flewelling March 4 th, 2009.
3D from multiple views : Rendering and Image Processing Alexei Efros …with a lot of slides stolen from Steve Seitz and Jianbo Shi.
CSCE 641 Computer Graphics: Image-based Modeling Jinxiang Chai.
COMP 290 Computer Vision - Spring Motion II - Estimation of Motion field / 3-D construction from motion Yongjik Kim.
CSE473/573 – Stereo Correspondence
Stereo Guest Lecture by Li Zhang
Project 1 artifact winners Project 2 questions Project 2 extra signup slots –Can take a second slot if you’d like Announcements.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2006 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Camera Calibration & Stereo Reconstruction Jinxiang Chai.
Stereo Matching Information Permeability For Stereo Matching – Cevahir Cigla and A.Aydın Alatan – Signal Processing: Image Communication, 2013 Radiometric.
Automatic Registration of Color Images to 3D Geometry Computer Graphics International 2009 Yunzhen Li and Kok-Lim Low School of Computing National University.
A Local Adaptive Approach for Dense Stereo Matching in Architectural Scene Reconstruction C. Stentoumis 1, L. Grammatikopoulos 2, I. Kalisperakis 2, E.
MESA LAB Multi-view image stitching Guimei Zhang MESA LAB MESA (Mechatronics, Embedded Systems and Automation) LAB School of Engineering, University of.
Recap from Monday Image Warping – Coordinate transforms – Linear transforms expressed in matrix form – Inverse transforms useful when synthesizing images.
WIEN Stereo-based Image and Video Analysis for Multimedia Applications Application: “Computer-generated Stereoscopic Paintings“ M. Gelautz, E. Stavrakis,
A Camera-Projector System for Real-Time 3D Video Marcelo Bernardes, Luiz Velho, Asla Sá, Paulo Carvalho IMPA - VISGRAF Laboratory Procams 2005.
CS654: Digital Image Analysis Lecture 8: Stereo Imaging.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
Stereo Readings Szeliski, Chapter 11 (through 11.5) Single image stereogram, by Niklas EenNiklas Een.
Stereo Many slides adapted from Steve Seitz.
Course 9 Texture. Definition: Texture is repeating patterns of local variations in image intensity, which is too fine to be distinguished. Texture evokes.
Project 2 code & artifact due Friday Midterm out tomorrow (check your ), due next Fri Announcements TexPoint fonts used in EMF. Read the TexPoint.
Stereo Many slides adapted from Steve Seitz. Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image image 1image.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Lecture 16: Stereo CS4670 / 5670: Computer Vision Noah Snavely Single image stereogram, by Niklas EenNiklas Een.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Robust and Accurate Surface Measurement Using Structured Light IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, VOL. 57, NO. 6, JUNE 2008 Rongqian.
Paper presentation topics 2. More on feature detection and descriptors 3. Shape and Matching 4. Indexing and Retrieval 5. More on 3D reconstruction 1.
Plan B (Exploiting Camera-Projector Homography). Methodology in Summary If the projection screen is flat, it is possible to directly establish the relationship.
Project 2 due today Project 3 out today Announcements TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAA.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Photoconsistency constraint C2 q C1 p l = 2 l = 3 Depth labels If this 3D point is visible in both cameras, pixels p and q should have similar intensities.
Energy minimization Another global approach to improve quality of correspondences Assumption: disparities vary (mostly) smoothly Minimize energy function:
Stereo CS4670 / 5670: Computer Vision Noah Snavely Single image stereogram, by Niklas EenNiklas Een.
AAFS 2004 Dallas Zeno Geradts
Coded Structured Light Coded Structured Light
A Plane-Based Approach to Mondrian Stereo Matching
Summary of “Efficient Deep Learning for Stereo Matching”
Semi-Global Matching with self-adjusting penalties
José Manuel Iñesta José Martínez Sotoca Mateo Buendía
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
DIGITAL SIGNAL PROCESSING
Approximate Models for Fast and Accurate Epipolar Geometry Estimation
SoC and FPGA Oriented High-quality Stereo Vision System
Common Classification Tasks
Machine Vision Acquisition of image data, followed by the processing and interpretation of these data by computer for some useful application like inspection,
Image Based Modeling and Rendering (PI: Malik)
Image Segmentation.
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Scalable light field coding using weighted binary images
Presentation transcript:

Presented by: Ali Agha March 02, 2009

Outline Sterevision overview Motivation & Contribution Structured light & method overview Related work Disparity computation Results Conclusion Future work

STEREO VISION When 3D information of a scene is needed

Depth from Disparity (x R -x L )/f = b/z disparity=d RL =(x R -x L )

Motivation of the presented paper “A taxonomy and evaluation of dense two-frame stereo correspondence algorithms”. Intl. J. Comp. Vis., Tsukuba Venus

need for more challenging scenes accurate ground truth information?? Motivation of the presented paper

Contributions of this work A method for acquiring high-complexity stereo image pairs with pixel-accurate correspondence information. Does not require the calibration of the light sources High resolution in comparison with range sensors

Process Pipeline This method uses structured light and consists of the following stages: Acquire all desired views under all illuminations. Rectify the images Decode the light patterns at each pixel to compute correspondences. Compute the view and illumination disparity and combine them

Structured light Structured-light techniques rely on projecting one or more special light patterns onto a scene, usually in order to directly acquire a range map of the scene

Structured light A pair of cameras and one or more light projectors are used

Related Work in Decoding light patterns Overview of Techniques Posdame et al Carrihill et al Boyer Kak LeMoigne et al Morita et al Vuylsteke et al Tajima et al Wust Capson Griffin et al Maruyama Abe J. Batlle, E. Mouaddib, and J. Salvi. Recent progress in coded structured light as a technique to solve the correspondence problem: a survey. Pat. Recog., 31(7):963–982, 1998.

Related work -CODED STRUCTURED LIGHT TECHNIQUES Posdamer-Daltschuler

Related work -CODED STRUCTURED LIGHT TECHNIQUES Inokuchi, Sato and Matsuda bits temporally Gray-coded pattern projection 8 bits temporally binary-coded pattern projection

Gray Code Using such binary images requires log 2 (n) patterns to distinguish among n locations.

Decoding the light patterns Using average of all-white and all-black In practice, the only reliable way is to project both the code pattern and its inverse. In surfaces with widely varying reflection properties, use two different exposure times (0.5 and 0.1 sec.). If this largest difference is still below a threshold, the pixel is labeled “unknown”

Disparity computation View disparities Illumination disparities Definition: views – the images taken by the cameras Illuminations – the structured light patterns projected onto the scene.

View disparities Assuming rectified views leads simple 1D search Practical issues: Occlusion Unknown code values (due to shadows or reflections). A perfect matching code value may not exist (interpolation errors) Several perfect matching code values may exist (limited resolution)

View disparities The first problem (partial occlusion) is unavoidable The number of unknown code values can be reduced by using more than one illumination source As a final consistency check, we establish disparities d LR and d RL independently and cross-check for consistency.

View disparities scene under illumination view disparities

Illumination disparities disparity between the cameras and the illumination sources. The difference in our case is that we can register these illumination disparities with our rectified view disparities d LR without the need to explicitly calibrate the illumination sources (video projectors).

Illumination disparities Relationship between the left view L and illum. source 0. Each pixel whose view disparity has been established can be considered a (homogeneous) 3D scene point S=[x,y,d,1] with projective depth d = d LR (x, y). The pixel’s illumination disparity (u 0L, v 0L ) P = M 0L Sin which P = [u 0L v 0L 1]

Practical Issues A small number of pixels with large disparity errors can strongly affect the least-squares fit. Outlier detection by iterating the above process. Only those pixels with low residual errors are selected as input to the next iteration.

Illumination disparities Given the projection matrix M 0L, we can now solve equation for d LR at all pixels Note that these disparities are available for all points illuminated by source 0, even those that are not visible from the right camera.

Combining the disparity estimates Remaining task is to combine the 2N + 2 disparity maps. Create combined maps for each of L and R separately Whenever there is a majority of values within close range, we use the average otherwise, the pixel is labeled unknown. L and R maps are checked for consistency, for unoccluded pixels, d LR (x, y) = − d RL (x + d LR (x, y), y),

Combined disparity Most stereo implementations work with much smaller image sizes. So, we downsample the images and disparity maps to quarter size (460 × 384). Note that for the downsampled images, we now have disparities with quarter-pixel accuracy.

Unknown Disparities A remaining issue is that of holes, i.e., unknown disparity values Small holes can be filled by interpolation Large holes may remain in areas where no illumination codes were available to begin with. Two main sources: surfaces that show very low reflection areas that are shadowed under all illuminations.

Results Two different scenes, Cones and Teddy.

Experiments In experimental setup, a single digital camera (Canon G1) translating on a linear stage, and one or two light projectors illuminating the scene from different directions.

Results

Verification To verify that stereo data sets are useful for evaluating stereo matching algorithms, several of the algorithms from the Middlebury Stereo Page has been ran on our new images.

Conclusion a new methodology to acquire highly precise and reliable ground truth disparity measurements camera-projector disparities, which can be used as an auxiliary source of information to increase the reliability of correspondences and to fill in missing data.

Considerations for Future work Exploiting in navigation Field of view is limited by the range of light projector Investigate the number of projected patterns which directly affect the speed of the method In daylight or dark places Invisible lights

Thank you Questions??

Related work -CODED STRUCTURED LIGHT TECHNIQUES Posdamer-Daltschuler

Related work -CODED STRUCTURED LIGHT TECHNIQUES Inokuchi, Sato and Matsuda bits temporally Gray-coded pattern projection 8 bits temporally binary-coded pattern projection

Related work -CODED STRUCTURED LIGHT TECHNIQUES Sato, Yamamoto and Inokuchi proposed to use a Liquid Crystal Device which allows an increased number of columns to be projected with a high accuracy. The system also improves the coded speed, against a slide projector, so the LCD can be electronically controlled.

If an object has a high textural contrast or any high reßected surface regions, then, some pattern segmentation errors can be produced. – Solution? The problem of a light projector is sometimes a result of heat irradiation onto the scene

Related work -CODED STRUCTURED LIGHT TECHNIQUES Hattori and Sato 1995 replace the light projector with a semiconductor laser, which gives a high power illumination with low heat irradiation. The proposed system, named Cubiscope The Cubiscope system

Related work – Carrihill-Hummelk Look at the notes

Related work – Boyer-Kak Colour

Related work – Le Moigne-Waxman not-coded grid patterns

Related work – Morita-Yakima-Sakata

Related work – Vuylsteke-Oosterlinck