Binocular Stereo Vision

Slides:



Advertisements
Similar presentations
The Primary Visual Cortex
Advertisements

Chapter 2.
Central Visual Processes. Anthony J Greene2 Central Visual Pathways I.Primary Visual Cortex Receptive Field Columns Hypercolumns II.Spatial Frequency.
What is vision Aristotle - vision is knowing what is where by looking.
COMPUTATIONAL NEUROSCIENCE FINAL PROJECT – DEPTH VISION Omri Perez 2013.
Anatomy/Physiology of Binocular Vision Goals –Follow the M and P pathway out of primary visual cortex –Answer where binocularly and disparity driven cells.
Read Pinker article for Thurs.. Seeing in Stereo.
Motion Depth Cues – Motion 1. Parallax. Motion Depth Cues – Parallax.
Binocular Disparity points nearer than horopter have crossed disparity
The visual system Lecture 1: Structure of the eye
Stereo matching “Stereo matching” is the correspondence problem –For a point in Image #1, where is the corresponding point in Image #2? C1C1 C2C2 ? ? C1C1.
1B50 – Percepts and Concepts Daniel J Hulme. Outline Cognitive Vision –Why do we want computers to see? –Why can’t computers see? –Introducing percepts.
Careers for Psychology and Neuroscience Majors Oct. 19th5-7pm in SU 300 Ballroom B.
Another viewpoint: V1 cells are spatial frequency filters
1 Computational Vision CSCI 363, Fall 2012 Lecture 26 Review for Exam 2.
Chapter 5 Human Stereopsis, Fusion, and Stereoscopic Virtual Environments.
1 Computational Vision CSCI 363, Fall 2012 Lecture 10 Spatial Frequency.
1 Computational Vision CSCI 363, Fall 2012 Lecture 20 Stereo, Motion.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
Two Views of Perception: Bottom-up processing: -Low-level -Feature driven Top-down processing -Knowledge driven -Unconscious inference (Helmholtz)
CS332 Visual Processing Department of Computer Science Wellesley College Binocular Stereo Vision Region-based stereo matching algorithms Properties of.
1 Computational Vision CSCI 363, Fall 2012 Lecture 5 The Retina.
1 Mid level vision, neglected yet still important Ken Nakayama Harvard University.
1 Perception and VR MONT 104S, Fall 2008 Lecture 4 Lightness, Brightness and Edges.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
1 Computational Vision CSCI 363, Fall 2012 Lecture 16 Stereopsis.
Grenoble Images Parole Signal Automatique Modeling of visual cortical processing to estimate binocular disparity Introduction - The objective is to estimate.
Understanding Psychophysics: Spatial Frequency & Contrast
Outline Of Today’s Discussion 1.LGN Projections & Color Opponency 2.Primary Visual Cortex: Structure 3.Primary Visual Cortex: Individual Cells.
1 Computational Vision CSCI 363, Fall 2012 Lecture 18 Stereopsis III.
Computational Vision CSCI 363, Fall 2012 Lecture 17 Stereopsis II
Independent Component Analysis features of Color & Stereo images Authors: Patrik O. Hoyer Aapo Hyvarinen CIS 526: Neural Computation Presented by: Ajay.
Processing visual information - pathways
Exploring Spatial Frequency Channels in Stereopsis
ORTH 140 NORMAL BINOCULAR SINGLE VISION AND MOTOR FUSION
Depth Perception, with Emphasis on Stereoscopic Vision
Computational Vision CSCI 363, Fall 2016 Lecture 15 Stereopsis
STEREOPSIS The Stereopsis Problem: Fusion and Reconstruction
EECS 274 Computer Vision Stereopsis.
Properties of human stereo processing
Common Classification Tasks
Space Perception and Binocular Vision
Why is Spatial Stereoacuity so Poor?
Early Processing in Biological Vision
Stereopsis: How the brain sees depth
“What Not” Detectors Help the Brain See in Depth
Binocular Stereo Vision
Spatial Vision (continued)
Binocular Stereo Vision
Binocular Stereo Vision
Christopher C. Pack, Richard T. Born, Margaret S. Livingstone  Neuron 
Binocular Stereo Vision
Binocular Stereo Vision
Active Vision Carol Colby Rebecca Berman Cathy Dunn Chris Genovese
Stereopsis Current Biology
Motion-Based Analysis of Spatial Patterns by the Human Visual System
Binocular Disparity and the Perception of Depth
Detecting image intensity changes
A Map for Horizontal Disparity in Monkey V2
Syed A. Chowdhury, Gregory C. DeAngelis  Neuron 
Course 6 Stereo.
Neural Mechanisms of Visual Motion Perception in Primates
Fangtu T. Qiu, Rüdiger von der Heydt  Neuron 
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Receptive Fields of Disparity-Tuned Simple Cells in Macaque V1
Binocular Stereo Vision
Stereoscopic Surface Perception
Outline Human Visual Information Processing – cont.
Detection of image intensity changes
Visual Perception: One World from Two Eyes
Presentation transcript:

Binocular Stereo Vision Properties of human stereo vision Marr-Poggio-Grimson multi-resolution stereo algorithm

Properties of human stereo processing Use features for stereo matching whose position and disparity can be measured very precisely Stereoacuity is only a few seconds of visual angle difference in depth  0.01 cm at a viewing distance of 30 cm

Properties of human stereo processing Matching features must appear similar in the left and right images For example, a left stereo image cannot be fused with a negative of the right image…

Properties of human stereo processing Only “fuse” objects within a limited range of depth around the fixation distance Vergence eye movements are needed to fuse objects over a larger range of depths

Properties of human stereo vision Human visual system can only tolerate small amounts of vertical disparity at a single eye position Vertical eye movements are needed to handle large vertical disparities

Properties of human stereo processing In the early stages of visual processing, the image is analyzed at multiple spatial scales … Hermann von Helmholtz coarse medium fine ... that play an important role in the solution to the stereo correspondence problem

Spatial frequency decomposition Any real signal, such as I(x), can be described as the sum of sinusoidal waves of different frequency, amplitude, and phase frequency e.g. cycles/deg I(x) Fourier Transform amplitude Two ways to describe 1D or 2D signals low frequency high

”Spatial frequency channels” in human vision Campbell & Robson, 1968

Spatial frequency channels

Properties of human stereo processing In the early stages of visual processing, the image is analyzed at multiple spatial scales… Hermann von Helmholtz • Stereo information at different scales can be processed independently • Stereo information at coarser scales can be ”fused” over a larger range of stereo disparity • Stereo information at coarser scales can trigger vergence eye movements that narrow the range of stereo disparity present in the images

Projection from the retina 1st cortical stage of visual processing: primary visual cortex (area V1) 1-11

Neural processing of stereo disparity monkey primary visual cortex

Neural mechanisms for stereo processing From G. Poggio & others: • neural recordings from monkey (area V1) • viewing random-dot stereograms zero disparity: at fixation distance near: in front of fixation distance far: behind fixation distance • (some) simple & complex cells in area V1 are selective for stereo disparity • neurons with large receptive fields are selective for a larger range of disparity ... but the stereo correspondence problem is not solved in V1!!

Selectivity for stereo boundaries in V2 Von der Heydt & colleagues: Some V2 cells are selective for the orientation, contrast, and side of border ownership of an edge ... for edges defined by luminance or stereo disparity “anti-correlated” stereogram Later, in area V4, neural responses to stereo disparity appear to correspond more closely to perceived depth

In summary, some key points… Image features used for matching: ~simple, precise locations, similar between left/right images At single fixation, match features over a limited range of horizontal & vertical disparity Eye movements used to match features over larger range of disparity Stereo matching performed at multiple scales ~independently, disparity range depends on scale Neurons selective for different ranges of stereo disparity, multiple processing stages V1  V2  V4

Matching features for the MPG stereo algorithm zero-crossings of image convolutions with 2G operators of different size L rough disparities over large range M accurate disparities over small range S

large w left large w right small w left small w right correct match outside search range at small scale

vergence eye movements! large w left right vergence eye movements! small w left right correct match now inside search range at small scale

Stereo images (Tsukuba, CMU)

Zero-crossings for stereo matching + - … …

Simplified MPG algorithm, Part 1 To determine initial correspondence: (1) Find zero-crossings using a 2G operator with central positive width w (2) For each horizontal slice: (2.1) Find the nearest neighbors in the right image for each zero-crossing fragment in the left image (2.2) Fine the nearest neighbors in the left image for each zero-crossing fragment in the right image (2.3) For each pair of zero-crossing fragments that are closest neighbors of one another, let the right fragment be separated by δinitial from the left. Determine whether δinitial is within the matching tolerance, m. If so, consider the zero-crossing fragments matched with disparity δinitial m = w/2

Simplified MPG algorithm, Part 2 To determine final correspondence: (1) Find zero-crossings using a 2G operator with reduced width w/2 (2) For each horizontal slice: (2.1) For each zero-crossing in the left image: (2.1.1) Determine the nearest zero-crossing fragment in the left image that matched when the 2G operator width was w (2.1.2) Offset the zero-crossing fragment by a distance δinitial, the disparity of the nearest matching zero-crossing fragment found at the lower resolution with operator width w (2.2) Find the nearest neighbors in the right image for each zero-crossing fragment in the left image (2.3) Fine the nearest neighbors in the left image for each zero-crossing fragment in the right image (2.4) For each pair of zero-crossing fragments that are closest neighbors of one another, let the right fragment be separated by δnew from the left. Determine whether δnew is within the reduced matching tolerance, m/2. If so, consider the zero-crossing fragments matched with disparity δfinal = δnew + δinitial

w = 8 m = 4 w = 4 m = 2 w = 4 m = 2 Coarse-scale zero-crossings: Use coarse-scale disparities to guide fine-scale matching: w = 4 m = 2 Ignore coarse-scale disparities: w = 4 m = 2 1-23 23