CS332 Visual Processing Department of Computer Science Wellesley College Binocular Stereo Vision Region-based stereo matching algorithms Properties of.

Slides:



Advertisements
Similar presentations
Chapter 2.
Advertisements

Central Visual Processes. Anthony J Greene2 Central Visual Pathways I.Primary Visual Cortex Receptive Field Columns Hypercolumns II.Spatial Frequency.
Seeing 3D from 2D Images. How to make a 2D image appear as 3D! ► Output and input is typically 2D Images ► Yet we want to show a 3D world! ► How can we.
CS 376b Introduction to Computer Vision 04 / 21 / 2008 Instructor: Michael Eckmann.
COMPUTATIONAL NEUROSCIENCE FINAL PROJECT – DEPTH VISION Omri Perez 2013.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
The Apparatus. Seeing in Stereo It’s very hard to read words if there are multiple images on your retina.
Lecture 4 Linear Filters and Convolution
Read Pinker article for Thurs.. Seeing in Stereo.
CS 376b Introduction to Computer Vision 02 / 27 / 2008 Instructor: Michael Eckmann.
Motion Depth Cues – Motion 1. Parallax. Motion Depth Cues – Parallax.
Fitting a Model to Data Reading: 15.1,
Binocular Disparity points nearer than horopter have crossed disparity
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
CSE473/573 – Stereo Correspondence
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
Stereo matching “Stereo matching” is the correspondence problem –For a point in Image #1, where is the corresponding point in Image #2? C1C1 C2C2 ? ? C1C1.
1B50 – Percepts and Concepts Daniel J Hulme. Outline Cognitive Vision –Why do we want computers to see? –Why can’t computers see? –Introducing percepts.
Careers for Psychology and Neuroscience Majors Oct. 19th5-7pm in SU 300 Ballroom B.
Divisors and Factors Section 2.3. Objectives Write a counting number as the product of two factors in all possible ways List all of the factors (divisors)
CS 376b Introduction to Computer Vision 02 / 26 / 2008 Instructor: Michael Eckmann.
1 Computational Vision CSCI 363, Fall 2012 Lecture 26 Review for Exam 2.
Chapter 5 Human Stereopsis, Fusion, and Stereoscopic Virtual Environments.
Integral University EC-024 Digital Image Processing.
Shape from Stereo  Disparity between two images  Photogrammetry  Finding Corresponding Points Correlation based methods Feature based methods.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
1 Computational Vision CSCI 363, Fall 2012 Lecture 20 Stereo, Motion.
Digital Image Processing CCS331 Relationships of Pixel 1.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
Experiments Irena Farberov,Andrey Terushkin. Stereo The word "stereo" comes from the Greek word "stereos" which means firm or solid. With stereo vision.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Stereo Viewing Mel Slater Virtual Environments
: Chapter 11: Three Dimensional Image Processing 1 Montri Karnjanadecha ac.th/~montri Image.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
CS 450: COMPUTER GRAPHICS PROJECTIONS SPRING 2015 DR. MICHAEL J. REALE.
1 Perception and VR MONT 104S, Fall 2008 Lecture 4 Lightness, Brightness and Edges.
CS654: Digital Image Analysis
1 Computational Vision CSCI 363, Fall 2012 Lecture 6 Edge Detection.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
1 Computational Vision CSCI 363, Fall 2012 Lecture 16 Stereopsis.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Photoconsistency constraint C2 q C1 p l = 2 l = 3 Depth labels If this 3D point is visible in both cameras, pixels p and q should have similar intensities.
1 Computational Vision CSCI 363, Fall 2012 Lecture 18 Stereopsis III.
Computational Vision CSCI 363, Fall 2012 Lecture 17 Stereopsis II
Independent Component Analysis features of Color & Stereo images Authors: Patrik O. Hoyer Aapo Hyvarinen CIS 526: Neural Computation Presented by: Ajay.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
Summary of “Efficient Deep Learning for Stereo Matching”
Computational Vision CSCI 363, Fall 2016 Lecture 15 Stereopsis
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Properties of human stereo processing
Common Classification Tasks
Early Processing in Biological Vision
Binocular Stereo Vision
Binocular Stereo Vision
Binocular Stereo Vision
Efficient Receptive Field Tiling in Primate V1
Binocular Stereo Vision
Binocular Stereo Vision
Binocular Stereo Vision
Binocular Disparity and the Perception of Depth
Detecting image intensity changes
Magnetic Resonance Imaging
Course 6 Stereo.
Binocular Stereo Vision
Receptive Fields of Disparity-Tuned Simple Cells in Macaque V1
Binocular Stereo Vision
Detection of image intensity changes
Efficient Receptive Field Tiling in Primate V1
Presentation transcript:

CS332 Visual Processing Department of Computer Science Wellesley College Binocular Stereo Vision Region-based stereo matching algorithms Properties of human stereo processing

Solving the stereo correspondence problem 1-2

1-3 (1) sum of absolute differences Σ | p left – p right | (2) normalized correlation (p left – p left ) (p right – p right ) patch σ p left σ p right Σ patch Measuring goodness of match between patches (1/n) Optional: divide by n = number of pixels in patch

Region-based stereo matching algorithm for each row r for each column c let p left be a square patch centered on (r,c) in the left image initialize best match score m best to ∞ initialize best disparity d best for each disparity d from – d range to + d range let p right be a square patch centered on (r,c+d) in the right image compute the match score m between p left and p right (sum of absolute differences) if (m < m best ), assign m best = m and d best = d record d best in the disparity map at (r,c) 1-4 (normalized correlation) How are the assumptions used??

1-5 leftright The real world works against us sometimes…

Example: Region-based stereo matching, using filtered images and sum-of-absolute differences 1-6 (from Carolyn Kim, 2013)

1-7 Properties of human stereo processing Use features for stereo matching whose position and disparity can be measured very precisely Stereoacuity is only a few seconds of visual angle difference in depth  0.01 cm at a viewing distance of 30 cm

1-8 Properties of human stereo processing Matching features must appear similar in the left and right images For example, we can’t fuse a left stereo image with a negative of the right image…

1-9 Properties of human stereo processing Only “fuse” objects within a limited range of depth around the fixation distance Vergence eye movements are needed to fuse objects over larger range of depths

1-10 Properties of human stereo vision We can only tolerate small amounts of vertical disparity at a single eye position Vertical eye movements are needed to handle large vertical disparities

1-11 Properties of human stereo processing In the early stages of visual processing, the image is analyzed at multiple spatial scales… Stereo information at multiple scales can be processed independently

1-12 Neural mechanisms for stereo processing G. Poggio & colleagues: complex cells in area V1 of primate visual cortex are selective for stereo disparity neurons that are selective for a larger disparity range have larger receptive fields zero disparity: at fixation distance near: in front of point of fixation far: behind point of fixation

1-13 In summary, some key points… Image features used for matching: simple, precise locations, multiple scales, similar between left/right images At single fixation position, match features over a limited range of horizontal & vertical disparity Eye movements used to match features over larger range of disparity Neural mechanisms selective for particular ranges of stereo disparity

1-14 Matching features for the MPG stereo algorithm zero-crossings of convolutions with  2 G operators of different size L M S rough disparities over large range accurate disparities over small range

1-15 large w left large w right small w left small w right correct match outside search range at small scale

1-16 large w left right small w left right correct match now inside search range at small scale vergence eye movements!

1-17 Stereo images (Tsukuba, CMU)

1-18 Zero-crossings for stereo matching - + ……

1-19 Simplified MPG algorithm, Part 1 To determine initial correspondence: (1) Find zero-crossings using a  2 G operator with central positive width w (2) For each horizontal slice: (2.1) Find the nearest neighbors in the right image for each zero-crossing fragment in the left image (2.2) Fine the nearest neighbors in the left image for each zero-crossing fragment in the right image (2.3) For each pair of zero-crossing fragments that are closest neighbors of one another, let the right fragment be separated by δ initial from the left. Determine whether δ initial is within the matching tolerance, m. If so, consider the zero-crossing fragments matched with disparity δ initial m = w/2

1-20 Simplified MPG algorithm, Part 2 To determine final correspondence: (1) Find zero-crossings using a  2 G operator with reduced width w/2 (2) For each horizontal slice: (2.1) For each zero-crossing in the left image: (2.1.1) Determine the nearest zero-crossing fragment in the left image that matched when the  2 G operator width was w (2.1.2) Offset the zero-crossing fragment by a distance δ initial, the disparity of the nearest matching zero-crossing fragment found at the lower resolution with operator width w (2.2) Find the nearest neighbors in the right image for each zero- crossing fragment in the left image (2.3) Fine the nearest neighbors in the left image for each zero- crossing fragment in the right image (2.4) For each pair of zero-crossing fragments that are closest neighbors of one another, let the right fragment be separated by δ new from the left. Determine whether δ new is within the reduced matching tolerance, m/2. If so, consider the zero-crossing fragments matched with disparity δ final = δ new + δ initial

1-21 Coarse-scale zero-crossings: Use coarse-scale disparities to guide fine-scale matching: Ignore coarse-scale disparities: w = 8 m = 4 w = 4 m = 2 w = 4 m = 2