December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.

Slides:



Advertisements
Similar presentations
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Advertisements

Dynamic Occlusion Analysis in Optical Flow Fields
Computer Vision Lecture 16: Texture
Computer Vision Lecture 16: Region Representation
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
CSE473/573 – Stereo Correspondence
Three-Dimensional Concepts
December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding.
Computer Vision Lecture 3: Digital Images
September 25, 2014Computer Vision Lecture 6: Spatial Filtering 1 Computing Object Orientation We compute the orientation of an object as the orientation.
Optical flow (motion vector) computation Course: Computer Graphics and Image Processing Semester:Fall 2002 Presenter:Nilesh Ghubade
Mohammed Rizwan Adil, Chidambaram Alagappan., and Swathi Dumpala Basaveswara.
Computer vision.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Last Week Recognized the fact that the 2D image is a representation of a 3D scene thus contains a consistent interpretation –Labeled edges –Labeled vertices.
1/20 Obtaining Shape from Scanning Electron Microscope Using Hopfield Neural Network Yuji Iwahori 1, Haruki Kawanaka 1, Shinji Fukui 2 and Kenji Funahashi.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
October 14, 2014Computer Vision Lecture 11: Image Segmentation I 1Contours How should we represent contours? A good contour representation should meet.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
September 23, 2014Computer Vision Lecture 5: Binary Image Processing 1 Binary Images Binary images are grayscale images with only two possible levels of.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
Digital Image Processing CCS331 Relationships of Pixel 1.
Computer Vision Why study Computer Vision? Images and movies are everywhere Fast-growing collection of useful applications –building representations.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
Course 9 Texture. Definition: Texture is repeating patterns of local variations in image intensity, which is too fine to be distinguished. Texture evokes.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Motion Analysis using Optical flow CIS750 Presentation Student: Wan Wang Prof: Longin Jan Latecki Spring 2003 CIS Dept of Temple.
1 Perception and VR MONT 104S, Fall 2008 Lecture 21 More Graphics for VR.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
3D Imaging Motion.
1 Self-Calibration and Neural Network Implementation of Photometric Stereo Yuji IWAHORI, Yumi WATANABE, Robert J. WOODHAM and Akira IWATA.
Autonomous Robots Vision © Manfred Huber 2014.
(c) 2000, 2001 SNU CSE Biointelligence Lab Finding Region Another method for processing image  to find “regions” Finding regions  Finding outlines.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
October 16, 2014Computer Vision Lecture 12: Image Segmentation II 1 Hough Transform The Hough transform is a very general technique for feature detection.
October 1, 2013Computer Vision Lecture 9: From Edges to Contours 1 Canny Edge Detector However, usually there will still be noise in the array E[i, j],
1Ellen L. Walker 3D Vision Why? The world is 3D Not all useful information is readily available in 2D Why so hard? “Inverse problem”: one image = many.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Digital Image Processing CSC331
September 26, 2013Computer Vision Lecture 8: Edge Detection II 1Gradient In the one-dimensional case, a step edge corresponds to a local peak in the first.
Computational Vision CSCI 363, Fall 2012 Lecture 17 Stereopsis II
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
MAN-522 Computer Vision Spring
Processing visual information for Computer Vision
Computational Vision CSCI 363, Fall 2016 Lecture 15 Stereopsis
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Mean Shift Segmentation
Fourier Transform: Real-World Images
Common Classification Tasks
Computer Vision Lecture 5: Binary Image Processing
Computer Vision Lecture 4: Color
Fitting Curve Models to Edges
Range Imaging Through Triangulation
Presented by: Cindy Yan EE6358 Computer Vision
Computer Vision Lecture 9: Edge Detection II
Computer Vision Lecture 16: Texture II
Magnetic Resonance Imaging
Course 6 Stereo.
Introduction to Artificial Intelligence Lecture 24: Computer Vision IV
Fourier Transform of Boundaries
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Intensity Transformation
Shape from Shading and Texture
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline distance b gives us a more precise estimate of depth z. However, the greater b, the more different are the two viewing angles, and the more difficult it can become to determine the correspondence between the two images. This brings us to the main problem in stereo vision: How can we find the conjugate pairs in our stereo images? This problem is called stereo matching.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 2 Stereo Matching In stereo matching, we have to solve a problem that is still under investigation by many researchers, called the correspondence problem. It can be phrased like this: For each point in the left image, find the corresponding point in the right image. The idea underlying all stereo matching algorithms is that these two points should be similar to each other. So we need a measure for similarity. Moreover, we need to find matchable features.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 3 Stereo Matching A straightforward approach to stereo matching uses pyramids, i.e., representations of the two camera images at various resolutions. The low-resolution versions of two corresponding rows are used to determine the “rough” matching, i.e. large patterns that match between the images. The precise disparity is determined in the high- resolution images. As a measure of similarity, we could simply use pieces of a row in the left image as convolution filters and apply them to the corresponding row in the right image.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 4 Generating Interesting Points Using interpolation to determine the depth of points within large homogeneous areas may cause large errors. To overcome this problem, we can generate additional interesting points that can be matched between the two images. The idea is to use structured light, i.e., project a pattern of light onto the visual scene. This creates additional variance in the brightness of pixels and increases the number of interesting points.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 5 Generating Interesting Points

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 6 Shape from Shading Besides binocular disparity, there are many different ways of depth estimation based on monocular information. For example, if we know the reflective properties of the surfaces in our scene and the position of the light source, we can use shape from shading techniques: Basically, since the amount of light reflected by a surface depends on its angle towards the light source, we can estimate the orientation of surfaces based on their intensity. More sophisticated methods also use the contours of shadows cast by objects to estimate the shape and orientation of those objects.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 7 Photometric Stereo To improve the coarse estimates of orientation derived from shape from shading methods, we can use photometric stereo. This technique uses three light sources that are located at different, known positions. Three images are taken, one for each light source, with the other two light sources being turned off. This way we determine three different intensities for each surface in the scene. These three values put narrow constraints on the possible orientation of a surface and allow a rather precise estimation.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 8 Shape from Texture As we discussed before, the texture gradient gives us valuable monocular depth information. At any point in the image showing texture, the texture gradient is a two-dimensional vector pointing towards the steepest increase in the size of texture elements. The texture gradient across a surface allows a good estimate of the spatial orientation of that surface. Of course, it is important for this technique that the image has high resolution and a precise method of texture size (granularity) measurement is used.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 9 Shape from Motion The shape from motion technique s similar to binocular stereo, but it uses only one camera. This camera is moved while it takes images from the visual scene. This way two images with a specific baseline distance can be obtained, and depth can be computed just like for binocular stereo. We can even use more than two images in this computation to get more robust measurements of depth. The disadvantages of shape from motion techniques are the technical overhead for moving the camera and the reduced temporal resolution.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 10 Range Imaging If we can determine the depth of every pixel in an image, we can make this information available in the form of a range image. A range image has exactly the same size and number of pixels as the original image. However, each pixel does not specify color or intensity, but the depth of that pixel in the original image, encoded as grayscale. Usually, the brighter a pixel in a range image is, the closer the corresponding pixel is to the observer. By providing both the original and the range image, we basically define a 3D image.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 11 Sample Range Image of a Mug

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 12 Range Imaging Through Triangulation How can we obtain precise depth information for every pixel in the image of a scene? One precise but slow method uses a laser that can rotate around its vertical axis and can also assume different vertical positions. This laser systematically and sequentially illuminates points in the image. A scene camera determines the position of every single point in its picture. The trick is that this camera looks at the scene from a different direction than does the laser pointer. Therefore, the depth of every point can be easily and precisely determined through triangulation:

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 13 Range Imaging Through Triangulation

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 14 Range Imaging Through Triangulation Obviously, this is a very slow process and not suitable for dynamic scenes. To speed things up, we can use a laser that projects a vertical line of light onto the scene. This laser rotates around its vertical axis and thereby moves the vertical line of light across the scene. Since only the horizontal positions of points vary and give us depth information, the vertical order of points is preserved. This allows us to compute the depth of each point along the vertical line without ambiguities.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 15 Range Imaging Through Triangulation

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 16 Range Imaging Through Triangulation Although this method is faster, it still requires a complete horizontal scan before a depth image is complete. Maybe we should use a pattern of many vertical lines that only needs to be shifted by the distance between neighboring lines? The disadvantage of this idea is that we could confuse points in different vertical lines, i.e., associate points with incorrect projection angles. However, we can overcome this problem by taking multiple images of the same scene with the pattern in the same position. In each picture, a different subset of lines is projected.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 17 Range Imaging Through Triangulation Then each line can be uniquely identified by its pattern of presence/absence across the images. For example, for 7 vertical lines we need a series of 3 images to do this encoding: Line #1 Line #2 Line #3 Line #4 Line #5 Line #6 Line #7 Image a offoffoffonononon Image b offononoffoffonon Image c onoffonoffonoffon Obviously, with this technique we can encode up to (n – 1) lines using log 2 (n) images. Therefore, this method is more efficient than the single-line scanning technique.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 18 Range Imaging Through Triangulation

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 19 Now we will talk about… Motion Analysis

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 20 Motion analysis Motion analysis is dealing with three main groups of motion- related problems: Motion detection Moving object detection and location. Derivation of 3D object properties. Motion analysis and object tracking combine two separate but inter-related components: Localization and representation of the object of interest (target). Trajectory filtering and data association. One or the other may be more important based on the nature of the motion application.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 21 Motion analysis

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 22 Differential Motion Analysis A simple method for motion detection is the subtraction of two or more images in a given image sequence. Usually, this method results in a difference image d(i, j), in which non-zero values indicate areas with motion. For given images f 1 and f 2, d(i, j) can be computed as follows: d(i, j) = 0 if | f 1 (i, j) – f 2 (i, j) |   = 1 otherwise

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 23 Differential Motion Analysis

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 24 Difference Pictures Another example of a difference picture that indicates the motion of objects (  = 25).

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 25 Difference Pictures Applying a size filter (size 10) to remove noise from a difference picture.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 26 Difference Pictures The differential method can rather easily be “tricked”. Here, the indicated changes were induced by changes in the illumination instead of object or camera motion (again,  = 25).

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 27 Differential Motion Analysis In order to determine the direction of motion, we can compute the cumulative difference image for a sequence f 1, …, f n of more than two images: Here, f 1 is used as the reference image, and the weight coefficients a k can be used to give greater weight to more recent frames and thereby highlight the current object positions.

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 28 Cumulative Difference Image Example: Sequence of 4 images: a 2 = a 3 = 2a 4 = 4 Result:

December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 29 Differential Motion Analysis Generally speaking, while differential motion analysis is well-suited for motion detection, it is not ideal for the analysis of motion characteristics.