Depth Estimation (Sz 11) “3d” movies are trendy lately, but arguably they should be called “stereo 3d” movies because movies already have many 3d cues.

Slides:



Advertisements
Similar presentations
Stereo Vision Reading: Chapter 11
Advertisements

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923.
Gratuitous Picture US Naval Artillery Rangefinder from World War I (1918)!!
Stereo Many slides adapted from Steve Seitz. Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image Where does the.
Recap from Previous Lecture Tone Mapping – Preserve local contrast or detail at the expense of large scale contrast. – Changing the brightness within.
Lecture 8: Stereo.
Stereo.
776 Computer Vision Jan-Michael Frahm, Enrique Dunn Spring 2012.
Camera calibration and epipolar geometry
CS6670: Computer Vision Noah Snavely Lecture 17: Stereo
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2005 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
Review Previous section: – Feature detection and matching – Model fitting and outlier rejection.
Stereo Binocular Stereo Calibration (finish up) Next Time Motivation
Stereopsis Mark Twain at Pool Table", no date, UCR Museum of Photography.
3D from multiple views : Rendering and Image Processing Alexei Efros …with a lot of slides stolen from Steve Seitz and Jianbo Shi.
Stereo and Structure from Motion
CSE473/573 – Stereo Correspondence
Announcements PS3 Due Thursday PS4 Available today, due 4/17. Quiz 2 4/24.
Stereo Guest Lecture by Li Zhang
Project 1 artifact winners Project 2 questions Project 2 extra signup slots –Can take a second slot if you’d like Announcements.
The Camera : Computational Photography Alexei Efros, CMU, Fall 2008.
Midterm went out on Tuesday (due next Tuesday) Project 3 out today Announcements.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2006 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
Project 4 Results Representation – SIFT and HoG are popular and successful. Data – Hugely varying results from hard mining. Learning – Non-linear classifier.
Wednesday March 23 Kristen Grauman UT-Austin
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Epipolar Geometry and Stereo Vision Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 04/12/11 Many slides adapted from Lana Lazebnik,
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
Stereo Readings Trucco & Verri, Chapter 7 –Read through 7.1, 7.2.1, 7.2.2, 7.3.1, 7.3.2, and 7.4, –The rest is optional. Single image stereogram,
Epipolar Geometry and Stereo Vision Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/05/15 Many slides adapted from Lana Lazebnik,
Camera Calibration & Stereo Reconstruction Jinxiang Chai.
Announcements Project 1 artifact winners Project 2 questions
Structure from images. Calibration Review: Pinhole Camera.
Recap from Wednesday Two strategies for realistic rendering capture real world data synthesize from bottom up Both have existed for 500 years. Both are.
Recap from Monday Image Warping – Coordinate transforms – Linear transforms expressed in matrix form – Inverse transforms useful when synthesizing images.
Stereo Vision Reading: Chapter 11 Stereo matching computes depth from two or more images Subproblems: –Calibrating camera positions. –Finding all corresponding.
Stereo Readings Szeliski, Chapter 11 (through 11.5) Single image stereogram, by Niklas EenNiklas Een.
Stereo Many slides adapted from Steve Seitz.
Project 2 code & artifact due Friday Midterm out tomorrow (check your ), due next Fri Announcements TexPoint fonts used in EMF. Read the TexPoint.
Stereo Many slides adapted from Steve Seitz. Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image image 1image.
Computer Vision, Robert Pless
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Peripheral drift illusion. Multiple views Hartley and Zisserman Lowe stereo vision structure from motion optical flow.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
Bahadir K. Gunturk1 Phase Correlation Bahadir K. Gunturk2 Phase Correlation Take cross correlation Take inverse Fourier transform  Location of the impulse.
Lecture 16: Stereo CS4670 / 5670: Computer Vision Noah Snavely Single image stereogram, by Niklas EenNiklas Een.
stereo Outline : Remind class of 3d geometry Introduction
Digital Image Processing
776 Computer Vision Jan-Michael Frahm Spring 2012.
Solving for Stereo Correspondence Many slides drawn from Lana Lazebnik, UIUC.
Project 2 due today Project 3 out today Announcements TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAA.
Project 3 code & artifact due Tuesday Final project proposals due noon Wed (by ) –One-page writeup (from project web page), specifying: »Your team.
Multiple View Geometry and Stereo. Overview Single camera geometry – Recap of Homogenous coordinates – Perspective projection model – Camera calibration.
CSE 185 Introduction to Computer Vision Stereo 2.
Stereo CS4670 / 5670: Computer Vision Noah Snavely Single image stereogram, by Niklas EenNiklas Een.
Noah Snavely, Zhengqi Li
Epipolar Geometry and Stereo Vision
Think-Pair-Share What visual or physiological cues help us to perceive 3D shape and depth?
제 5 장 스테레오.
The Camera : Computational Photography
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Stereo and Structure from Motion
Announcements Midterms graded (handed back at end of lecture)
Thanks to Richard Szeliski and George Bebis for the use of some slides
Lecture 13: Cameras and geometry
The Camera : Computational Photography
Multiple View Geometry for Robotics
Announcements Midterm out today Project 1 demos.
Stereo vision Many slides adapted from Steve Seitz.
Presentation transcript:

Depth Estimation (Sz 11) “3d” movies are trendy lately, but arguably they should be called “stereo 3d” movies because movies already have many 3d cues. Slides from Steve Seitz, Robert Collins, James Hays

Multiple views Multi-view geometry, matching, invariant features, stereo vision Lowe Hartley and Zisserman CS 376 Lecture 16: Stereo 1

Why multiple views? Structure and depth are inherently ambiguous from single views. Images from Lana Lazebnik CS 376 Lecture 16: Stereo 1

Why multiple views? Structure and depth are inherently ambiguous from single views. P1 P2 P1’=P2’ Optical center CS 376 Lecture 16: Stereo 1

What cues help us to perceive depth (and 3D shape in general)? Think-Pair-Share What cues help us to perceive depth (and 3D shape in general)?

Shading [Figure from Prados & Faugeras 2006] CS 376 Lecture 16: Stereo 1 8

Focus/defocus Images from same point of view, different camera parameters 3d shape / depth estimates [figs from H. Jin and P. Favaro, 2002] CS 376 Lecture 16: Stereo 1

Texture CS 376 Lecture 16: Stereo 1 10 [From A.M. Loh. The recovery of 3-D structure using visual texture patterns. PhD thesis] CS 376 Lecture 16: Stereo 1 10

Perspective effects Image credit: S. Seitz

Motion CS 376 Lecture 16: Stereo 1 Figures from L. Zhang http://www.brainconnection.com/teasers/?main=illusion/motion-shape CS 376 Lecture 16: Stereo 1

Stereo vision Two cameras, simultaneous views Single moving camera and static scene

Stereo Vision Not that important for humans, especially at longer distances. Perhaps 10% of people are stereo blind. Many animals don’t have much stereo overlap in their fields of view

Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923

Teesta suspension bridge-Darjeeling, India

Woman getting eye exam during immigration procedure at Ellis Island, c Woman getting eye exam during immigration procedure at Ellis Island, c. 1905 - 1920 , UCR Museum of Phography

Mark Twain at Pool Table", no date, UCR Museum of Photography

Yes, you can be stereoblind. Ask audience to put finger tip in line with dot in vision, then ask them to focus on dot -> should see two fingers. Then, ask them to focus on finger -> should see two dots. If people only see one finger and one dot, then they might be stereoblind….. Is this true? CS 376 Lecture 16: Stereo 1

Random dot stereograms Julesz 1960: Do we identify local brightness patterns before fusion (monocular process) or after (binocular)? Think Pair Share – yes / no? how to test? CS 376 Lecture 16: Stereo 1 25

Random dot stereograms Julesz 1960: Do we identify local brightness patterns before fusion (monocular process) or after (binocular)? To test: pair of synthetic images obtained by randomly spraying black dots on white objects CS 376 Lecture 16: Stereo 1 26

Random dot stereograms Forsyth & Ponce CS 376 Lecture 16: Stereo 1 27

Random dot stereograms CS 376 Lecture 16: Stereo 1 28

2. Select a region in one image. 1. Create an image of suitable size. Fill it with random dots. Duplicate the image. 2. Select a region in one image. 3. Shift this region horizontally by a small amount. The stereogram is complete. CC BY-SA 3.0, https://en.wikipedia.org/wiki/Random_dot_stereogram

Random dot stereograms When viewed monocularly, they appear random; when viewed stereoscopically, see 3d structure. Human binocular fusion not directly associated with the physical retinas; must involve the central nervous system (V2, for instance). Imaginary “cyclopean retina” that combines the left and right image stimuli as a single unit. High level scene understanding not required for stereo…but, high level scene understanding is arguably better than stereo. According to Ralph Siegel, Dr. Bela Julesz had "unambiguously demonstrated that stereoscopic depth could be computed in the absence of any identifiable objects, in the absence of any perspective, in the absence of any cues available to either eye alone." Artists are more often stereo blind than average people. Being stereo blind may train you to be better at recognizing and thus depicting other monocular depth cues. CS 376 Lecture 16: Stereo 1 30

Stereo attention is weird wrt. mind’s eye - In the following figure, a viewer perceives an image which is a superposition of two images, one shown to each of the two eyes using the equivalent of spectacles for watching 3D movies. - To the viewer, it is as if the perceived image is shown simultaneously to both eyes --- we see a superposition of the two images, and the left eye image ‘fills the space’ in the right eye image; the ocular singleton appears identical to all the other background bars, i.e. we are blind to its distinctiveness. The uniquely tilted bar appears most distinctive from the background. Nevertheless, the ocular singleton often attracts attention more strongly than the orientation singleton (so that the first gaze shift is more frequently directed to the ocular rather than the orientation singleton) even when the viewer is told to find the latter as soon as possible and ignore all distractions. It is as if the ocular singleton is uniquely coloured and distracting, like the lone-red flower in a green field, except that we are “colour-blind” to it. Many vision scientists find this hard to believe without experiencing it themselves. Fig 5.9 from Li Zhaoping, Understanding Vision: Theory Models, and Data [Li Zhaoping, Understanding Vision] CS 376 Lecture 16: Stereo 1

Autostereograms – ‘Magic Eye’ Exploit disparity as depth cue using single image. (Single image random dot stereogram, Single image stereogram) Images from magiceye.com CS 376 Lecture 16: Stereo 1 32

Autostereograms Images from magiceye.com CS 376 Lecture 16: Stereo 1 33

Stereo Depth Estimation Slides by Kristen Grauman CS 376 Lecture 16: Stereo 1

Geometry for a simple stereo system First, assuming parallel optical axes, known camera parameters (i.e., calibrated cameras): Two cameras, simultaneous views CS 376 Lecture 16: Stereo 1

optical center (right) optical center (left) World point Depth of p image point (left) image point (right) Focal length optical center (right) optical center (left) baseline CS 376 Lecture 16: Stereo 1 40

Geometry for a simple stereo system Assume parallel optical axes, known camera parameters (i.e., calibrated cameras). What is expression for Z? Similar triangles (pl, P, pr) and (Ol, P, Or): disparity CS 376 Lecture 16: Stereo 1 41

Depth from disparity (x´,y´)=(x+D(x,y), y) image I(x,y) Disparity map D(x,y) image I´(x´,y´) (x´,y´)=(x+D(x,y), y) So if we could find the corresponding points in two images, we could estimate relative depth… CS 376 Lecture 16: Stereo 1

Basic stereo matching algorithm Rectify the two stereo images to transform epipolar lines into scanlines For each pixel x in the first image Find corresponding epipolar scanline in the right image Examine all pixels on the scanline and pick the best match x’ Use individual pixels? Use patches! This should look familiar... Compute disparity x-x’ and set depth(x) = fB/(x-x’)

Correspondence problem Intensity profiles Source: Andrew Zisserman

Correspondence problem Neighborhoods of corresponding points are similar in intensity patterns. Source: Andrew Zisserman

Correlation-based window matching Let’s take a particular window size placed in the left window and slide it on the right. Here you see the target in the left band

Correlation-based window matching This is the epipolar corresponding right band. When I slide that window across I get this…

Correlation-based window matching Using cross correlation, [a] you can see a clear peak at the right spot.

Correlation-based window matching But what about this less textured region?

Correlation-based window matching ??? We get this corrleation. [a] And where the peak is located is hard to tell. How could we fix that? Well… Textureless regions are non-distinct; high ambiguity for matches.

When will this work/not work? Think-Pair-Share When will this work/not work?

Stereo reconstruction pipeline What will cause errors? Camera calibration errors Poor image resolution Occlusions Violations of brightness constancy (specular reflections) Large motions Low-contrast image regions

Failures of correspondence search Occlusions, repetition Textureless surfaces Non-Lambertian surfaces, specularities

Effect of window size W = 3 W = 20 Smaller window Larger window + More detail More noise Larger window + Smoother disparity maps Less detail smaller window: more detail, more noise bigger window: less noise, more detail Adaptive window approach T. Kanade and M. Okutomi, A Stereo Matching Algorithm with an Adaptive Window: Theory and Experiment,, Proc. International Conference on Robotics and Automation, 1991. D. Scharstein and R. Szeliski. Stereo matching with nonlinear diffusion. International Journal of Computer Vision, 28(2):155-174, July 1998

Stereo results Scene Ground truth Data from University of Tsukuba Similar results on other images without ground truth Scene Ground truth

Results with window search Data Window-based matching Ground truth

Correspondence problem There are “soft” constraints to help identify corresponding points Similarity Disparity gradient – depth doesn’t change too quickly. Uniqueness Ordering CS 376 Lecture 17 Stereo

Uniqueness constraint Up to one match in right image for every point in left image Figure from Gee & Cipolla 1999 CS 376 Lecture 17 Stereo

Problem: Occlusion Uniqueness says “up to match” per pixel When is there no match? Occluded pixels

Disparity gradient constraint Assume piecewise continuous surface, so want disparity estimates to be locally smooth Figure from Gee & Cipolla 1999 CS 376 Lecture 17 Stereo

Ordering constraint Points on same surface (opaque object) will be in same order in both views Figure from Gee & Cipolla 1999 CS 376 Lecture 17 Stereo

Ordering constraint Won’t always hold, e.g. consider transparent object, or an occluding surface Figures from Forsyth & Ponce CS 376 Lecture 17 Stereo

Stereo as energy minimization What defines a good stereo correspondence? Match quality Want each pixel to find a good match in the other image Smoothness If two pixels are adjacent, they should (usually) move about the same amount

Scanline stereo Try to coherently match pixels on the entire scanline Different scanlines are still optimized independently Left image Right image intensity

“Shortest paths” for scan-line stereo Left image Right image Right occlusion s q Left occlusion t p correspondence Can be implemented with dynamic programming Ohta & Kanade ’85, Cox et al. ‘96 Slide credit: Y. Boykov

Coherent stereo on 2D grid Scanline stereo generates streaking artifacts Can’t use dynamic programming to find spatially coherent disparities/ correspondences on a 2D grid Patch-based Per scanline dynamic programming

Stereo matching as energy minimization D W1(i ) W2(i+D(i )) D(i ) data term smoothness term Energy functions of this form can be minimized using graph cuts Y. Boykov, O. Veksler, and R. Zabih, Fast Approximate Energy Minimization via Graph Cuts, PAMI 2001

Stereo matching as energy minimization D W1(i ) W2(i+D(i )) D(i ) Energy functions of this form can be minimized using graph cuts. Y. Boykov, O. Veksler, and R. Zabih, Fast Approximate Energy Minimization via Graph Cuts, PAMI 2001 Source: Steve Seitz CS 376 Lecture 17 Stereo

Constraints encoded in an energy function and solved using graph cuts Patch-based Graph cuts Ground truth Y. Boykov, O. Veksler, and R. Zabih, Fast Approximate Energy Minimization via Graph Cuts, PAMI 2001 For the latest and greatest: http://www.middlebury.edu/stereo/

Depth from disparity f x x’ baseline z C C’ X

Active stereo: project structured light onto object Li Zhang’s one-shot stereo camera 2 camera 1 projector camera 1 projector Simplifies the correspondence problem L. Zhang, B. Curless, and S. M. Seitz. Rapid Shape Acquisition Using Color Structured Light and Multi-pass Dynamic Programming. 3DPVT 2002

Active stereo with structured light Sequence of known expected appearance across surface

How does a depth camera work? Microsoft Kinect v1 Intel laptop depth camera

Kinect: Structured infrared light http://bbzippo.wordpress.com/2010/11/28/kinect-in-infrared/

How does a depth camera work? Stereo in infrared.

Digital Michelangelo Project Laser scanning Digital Michelangelo Project http://graphics.stanford.edu/projects/mich/ Optical triangulation Project a single stripe of laser light Scan it across the surface of the object This is a very precise version of structured light scanning

The Digital Michelangelo Project, Levoy et al. Laser scanned models The Digital Michelangelo Project, Levoy et al.

The Digital Michelangelo Project, Levoy et al. Laser scanned models The Digital Michelangelo Project, Levoy et al.

The Digital Michelangelo Project, Levoy et al. Laser scanned models The Digital Michelangelo Project, Levoy et al.

The Digital Michelangelo Project, Levoy et al. Laser scanned models The Digital Michelangelo Project, Levoy et al.

Time of Flight (Kinect V2) Depth cameras in HoloLens use time of flight “SONAR for light” Emit light of a known wavelength, and time how long it takes for it to come back

Time of Flight Humans can learn to echolocate!

Break here.

The Camera Many slides by Alexei A. Efros CS 129: Computational Photography James Hays, Brown, Spring 2011

How do we see the world? Let’s design a camera Idea 1: put a piece of film in front of an object Do we get a reasonable image? Slide by Steve Seitz

Pinhole camera Add a barrier to block off most of the rays This reduces blurring The opening known as the aperture How does this transform the image? It gets inverted Slide by Steve Seitz

Pinhole camera model Pinhole model: Captures pencil of rays – all rays through a single point The point is called Center of Projection (COP) The image is formed on the Image Plane Effective focal length f is distance from COP to Image Plane Slide by Steve Seitz

Dimensionality Reduction Machine (3D to 2D) 3D world 2D image What have we lost? Angles Depth, lengths Figures © Stephen E. Palmer, 2002

Funny things happen…

Lengths can’t be trusted... Figure by David Forsyth

…but humans adopt! We don’t make measurements in the image plane Müller-Lyer Illusion We don’t make measurements in the image plane http://www.michaelbach.de/ot/sze_muelue/index.html

Modeling projection The coordinate system We will use the pin-hole model as an approximation Put the optical center (Center Of Projection) at the origin Put the image plane (Projection Plane) in front of the COP Why? The camera looks down the negative z axis we need this if we want right-handed-coordinates This way the image is right-side-up Slide by Steve Seitz

Modeling projection Projection equations Compute intersection with PP of ray from (x,y,z) to COP Derived using similar triangles We get the projection by throwing out the last coordinate: Slide by Steve Seitz

Homogeneous coordinates Is this a linear transformation? no—division by z is nonlinear Trick: add one more coordinate: homogeneous image coordinates homogeneous scene coordinates Converting from homogeneous coordinates Slide by Steve Seitz

Perspective Projection Projection is a matrix multiply using homogeneous coordinates: divide by third coordinate This is known as perspective projection The matrix is the projection matrix Can also formulate as a 4x4 divide by fourth coordinate Slide by Steve Seitz

Orthographic Projection Special case of perspective projection Distance from the COP to the PP is infinite Also called “parallel projection” What’s the projection matrix? Image World Slide by Steve Seitz

Camera parameters Camera frame 2 Camera frame 1 Extrinsic parameters: Camera frame 1  Camera frame 2 Intrinsic parameters: Image coordinates relative to camera  Pixel coordinates Extrinsic params: rotation matrix and translation vector Intrinsic params: focal length, pixel sizes (mm), image center point, radial distortion parameters We’ll assume for now that these parameters are given and fixed. CS 376 Lecture 16: Stereo 1 107

Estimating depth with stereo Stereo: shape from “motion” between two views We’ll need to consider: Info on camera pose (“calibration”) Image point correspondences scene point image plane optical center CS 376 Lecture 16: Stereo 1 108

General case, with calibrated cameras The two cameras need not have parallel optical axes. Vs. CS 376 Lecture 16: Stereo 1

Stereo correspondence constraints Given p in left image, where can corresponding point p’ be? CS 376 Lecture 16: Stereo 1

Stereo correspondence constraints CS 376 Lecture 16: Stereo 1

Epipolar constraint Geometry of two views constrains where the corresponding pixel for some image point in the first view must occur in the second view. It must be on the line carved out by a plane connecting the world point and optical centers. CS 376 Lecture 16: Stereo 1 112

Epipolar geometry Epipolar Line Epipolar Plane Epipole Baseline http://www.ai.sri.com/~luong/research/Meta3DViewer/EpipolarGeo.html CS 376 Lecture 16: Stereo 1 113

Epipolar geometry: terms Baseline: line joining the camera centers Epipole: point of intersection of baseline with image plane Epipolar plane: plane containing baseline and world point Epipolar line: intersection of epipolar plane with the image plane All epipolar lines intersect at the epipole An epipolar plane intersects the left and right image planes in epipolar lines Why is the epipolar constraint useful? CS 376 Lecture 16: Stereo 1

Epipolar constraint This is useful because it reduces the correspondence problem to a 1D search along an epipolar line. Image from Andrew Zisserman

Example CS 376 Lecture 16: Stereo 1

Think Pair Share = camera center Where are the epipoles? What do the epipolar lines look like? X X a) b) X X c) d)

Example: converging cameras Figure from Hartley & Zisserman CS 376 Lecture 16: Stereo 1 118

Example: parallel cameras Where are the epipoles? Figure from Hartley & Zisserman CS 376 Lecture 16: Stereo 1 119

Example: Forward motion Epipole has same coordinates in both images. Points move along lines radiating from e: “Focus of expansion”

Example: Forward motion Epipole has same coordinates in both images. Points move along lines radiating from e: “Focus of expansion”

Fundamental matrix 𝑙 ′ =𝐹𝑥 𝑙= 𝐹 𝑇 𝑥′ 𝑥 ′ 𝐹𝑥=0 Let x be a point in left image, x’ in right image Epipolar relation x maps to epipolar line l’ x’ maps to epipolar line l Epipolar mapping described by a 3x3 matrix F: It follows that: l’ l x x’ 𝑙 ′ =𝐹𝑥 𝑙= 𝐹 𝑇 𝑥′ 𝑥 ′ 𝐹𝑥=0

Fundamental matrix 𝑥 ′ 𝐹𝑥=0 This matrix F is called the “Essential Matrix” when image intrinsic parameters are known the “Fundamental Matrix” more generally (uncalibrated case) Can solve for F from point correspondences Each (x, x’) pair gives one linear equation in entries of F F has 9 entries, but really only 7 degrees of freedom. With 8 points it is simple to solve for F, but it is also possible with 7. See Marc Pollefey’s notes for a nice tutorial 𝑥 ′ 𝐹𝑥=0

Stereo image rectification

Stereo image rectification Reproject image planes onto a common plane parallel to the line between camera centers Pixel motion is horizontal after this transformation Two homographies (3x3 transform), one for each input image reprojection C. Loop and Z. Zhang. Computing Rectifying Homographies for Stereo Vision. IEEE Conf. Computer Vision and Pattern Recognition, 1999.

Rectification example

Summary: Key idea: Epipolar constraint X X X x x’ x’ x’ Potential matches for x have to lie on the corresponding line l’. Potential matches for x’ have to lie on the corresponding line l.

Summary Epipolar geometry Stereo depth estimation Epipoles are intersection of baseline with image planes Matching point in second image is on a line passing through its epipole Fundamental matrix maps from a point in one image to a line (its epipolar line) in the other Can solve for F given corresponding points (e.g., interest points) Stereo depth estimation Estimate disparity by finding corresponding points along scanlines Depth is inverse to disparity

Stereo reconstruction pipeline Steps Calibrate cameras Rectify images Compute disparity Estimate depth