A Robust Super Resolution Method for Images of 3D Scenes Pablo L. Sala Department of Computer Science University of Toronto.

Slides:



Advertisements
Similar presentations
Vanishing points  .
Advertisements

Automatic Photo Pop-up Derek Hoiem Alexei A.Efros Martial Hebert Carnegie Mellon University.
DTAM: Dense Tracking and Mapping in Real-Time
3D Model Matching with Viewpoint-Invariant Patches(VIP) Reporter :鄒嘉恆 Date : 10/06/2009.
Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam Road, Hong Kong Three-dimensional curve reconstruction from.
July 27, 2002 Image Processing for K.R. Precision1 Image Processing Training Lecture 1 by Suthep Madarasmi, Ph.D. Assistant Professor Department of Computer.
A Sampled Texture Prior for Image Super-Resolution Lyndsey C. Pickup, Stephen J. Roberts and Andrew Zisserman, Robotics Research Group, University of Oxford.
Mapping: Scaling Rotation Translation Warp
Proportion Priors for Image Sequence Segmentation Claudia Nieuwenhuis, etc. ICCV 2013 Oral.
Shape From Light Field meets Robust PCA
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
Active Calibration of Cameras: Theory and Implementation Anup Basu Sung Huh CPSC 643 Individual Presentation II March 4 th,
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Bayesian Image Super-resolution, Continued Lyndsey C. Pickup, David P. Capel, Stephen J. Roberts and Andrew Zisserman, Robotics Research Group, University.
Last Time Pinhole camera model, projection
Motion Analysis Slides are from RPI Registration Class.
Motion Analysis (contd.) Slides are from RPI Registration Class.
Image Quilting for Texture Synthesis and Transfer Alexei A. Efros1,2 William T. Freeman2.
Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002.
ON THE IMPROVEMENT OF IMAGE REGISTRATION FOR HIGH ACCURACY SUPER-RESOLUTION Michalis Vrigkas, Christophoros Nikou, Lisimachos P. Kondi University of Ioannina.
Multiple-view Reconstruction from Points and Lines
Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Interactive Visual Media Group Microsoft Research Dept. of Computer.
Lecture 22: Structure from motion CS6670: Computer Vision Noah Snavely.
The plan for today Camera matrix
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
Structure Computation. How to compute the position of a point in 3- space given its image in two views and the camera matrices of those two views Use.
David Luebke Modeling and Rendering Architecture from Photographs A hybrid geometry- and image-based approach Debevec, Taylor, and Malik SIGGRAPH.
Path-Based Constraints for Accurate Scene Reconstruction from Aerial Video Mauricio Hess-Flores 1, Mark A. Duchaineau 2, Kenneth I. Joy 3 Abstract - This.
What Does the Scene Look Like From a Scene Point? Donald Tanguay August 7, 2002 M. Irani, T. Hassner, and P. Anandan ECCV 2002.
Geometry and Algebra of Multiple Views
Automatic Registration of Color Images to 3D Geometry Computer Graphics International 2009 Yunzhen Li and Kok-Lim Low School of Computing National University.
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
Y. Moses 11 Combining Photometric and Geometric Constraints Yael Moses IDC, Herzliya Joint work with Ilan Shimshoni and Michael Lindenbaum, the Technion.
Reconstructing 3D mesh from video image sequences supervisor : Mgr. Martin Samuelčik by Martin Bujňák specifications Master thesis
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Multiview Geometry and Stereopsis. Inputs: two images of a scene (taken from 2 viewpoints). Output: Depth map. Inputs: multiple images of a scene. Output:
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
1 Formation et Analyse d’Images Session 7 Daniela Hall 25 November 2004.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Matching Users and Items Across Domains to Improve the Recommendation Quality Created by: Chung-Yi Li, Shou-De Lin Presented by: I Gde Dharma Nugraha 1.
Vehicle Segmentation and Tracking From a Low-Angle Off-Axis Camera Neeraj K. Kanhere Committee members Dr. Stanley Birchfield Dr. Robert Schalkoff Dr.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
18 th August 2006 International Conference on Pattern Recognition 2006 Epipolar Geometry from Two Correspondences Michal Perďoch, Jiří Matas, Ondřej Chum.
Ch. 3: Geometric Camera Calibration
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
Feature Matching. Feature Space Outlier Rejection.
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Newton's method Wikpedia page
Yizhou Yu Texture-Mapping Real Scenes from Photographs Yizhou Yu Computer Science Division University of California at Berkeley Yizhou Yu Computer Science.
3D Reconstruction Using Image Sequence
Real-Time Dynamic Shadow Algorithms Evan Closson CSE 528.
Notes on HW 1 grading I gave full credit as long as you gave a description, confusion matrix, and working code Many people’s descriptions were quite short.
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Newton's method Wikpedia page
Media Processor Lab. Media Processor Lab. Trellis-based Parallel Stereo Matching Media Processor Lab. Sejong univ.
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
A Plane-Based Approach to Mondrian Stereo Matching
- photometric aspects of image formation gray level images
José Manuel Iñesta José Martínez Sotoca Mateo Buendía
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Motion and Optical Flow
Mauricio Hess-Flores1, Mark A. Duchaineau2, Kenneth I. Joy3
Common Classification Tasks
Vehicle Segmentation and Tracking in the Presence of Occlusions
A guide to SR different approaches
Announcements more panorama slots available now
Announcements more panorama slots available now
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Presentation transcript:

A Robust Super Resolution Method for Images of 3D Scenes Pablo L. Sala Department of Computer Science University of Toronto

Super Resolution “The process of obtaining, from a set of images that overlap, a new image that in the regions of overlapping has a higher resolution than that of each individual image.” This work: Introduces a method for robust super resolution Applies it to obtain higher resolution images of a 3D scene from a set of calibrated low-resolution images under the assumption that the scene can be approximated by planes lying in 3D space.

The Registration Problem The algorithm presented here starts from a discreet set of planes in 3D space and for each plane it back-projects all the input images onto that plane and performs super- resolution. Only the image pixels of 3D points of the scene that lie on the plane will be back- projected to the same points in the plane.

Problem Formulation Assumptions: The scene can be thought of as mostly lying on planes in 3D space. All viewpoints are located in a region separable by a plane from the scene, so to hold to a notion of scene depth. The Idea: A set of N calibrated images of the scene. A list of planes ordered by its depth. Images set is partitioned in disjoint subsets of relatively same size. Super-resolution attempted on each plane using the back-projection of the images of each subset on that plane. The obtained higher resolution images are compared. The regions of coincidence will be assumed to correspond to parts of the scene lying on the plane. The portions of each image that correspond to these regions will not be taken into account when applying super-resolution on the following planes.

Standard Super-Resolution Super-resolution can be formulated as an optimization problem: X is the unknown higher resolution image x j are the low resolution images F j = D j H j W j are the imaging formation matrices, with D j a decimation, H j a blurring, and W j a geometric warp matrix. Deriving E with respect to X and equaling it to 0 we get to:

Robust Super-Resolution Objective function: Error: Cauchy’s robust estimator:

Robust Super-Resolution Deriving E with respect to X: where W j is a diagonal matrix such that: Iteratively re-weighted least squares: Initial guess X 0 : the higher resolution image computed using the standard version of super-resolution.

The Algorithm Input: L a list of planes in 3D space; I j images; P j camera matrices Set O =  (set of the occluding 3D points) Randomly partition the images set into two or more disjoint subsets S 1,..., S R. For each plane  in L do: For each subset S i do: Use , P j and O to compute the image formation matrices F j X 0 = S-Resolution of images in S i Y i = Robust S-R of images in S i, using X 0 as the initial guess R = region R of  where all Y i coincide Assume R is the portion of the scene that lies on  O = O  R

Observations For the algorithm to be effective: The 3D planes uniformly distributed, close one to another, and covering all scene depth, so to detect all scene portions to be aware of occlusions. The scene should be highly textured for the similarity check to work well.

Experimental Results Synthetic scene:

12 images were used as input 2 disjoint subsets Experimental Results

Standard vs Robust Super-Resolution: Experimental Results

Output images: Experimental Results

Results get worse as we advance in the 3D planes due to errors in defining precisely what portion of the scene lies in each plane Using more images and partitioning the images set in more than just two subsets will definitely give more accuracy to the similarity check. Using color images instead of just B&W. The super- resolution steps of the algorithm would be applied to each color channel separately and similarity would be assumed when the difference in reconstruction is close to zero for all three channels. Observations and Future Work