Mauricio Hess-Flores1, Mark A. Duchaineau2, Kenneth I. Joy3

Slides:



Advertisements
Similar presentations
Vanishing points  .
Advertisements

Practical Camera Auto-Calibration Based on Object Appearance and Motion for Traffic Scene Visual Surveillance Zhaoxiang Zhang, Min Li, Kaiqi Huang and.
A Robust Super Resolution Method for Images of 3D Scenes Pablo L. Sala Department of Computer Science University of Toronto.
3D Model Matching with Viewpoint-Invariant Patches(VIP) Reporter :鄒嘉恆 Date : 10/06/2009.
The fundamental matrix F
CSE473/573 – Stereo and Multiple View Geometry
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Multiview Reconstruction. Why More Than 2 Views? BaselineBaseline – Too short – low accuracy – Too long – matching becomes hard.
Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam Road, Hong Kong Three-dimensional curve reconstruction from.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Tracking Multiple Occluding People by Localizing on Multiple Scene Planes Saad M. Khan and Mubarak Shah, PAMI, VOL. 31, NO. 3, MARCH 2009, Donguk Seo
Active Calibration of Cameras: Theory and Implementation Anup Basu Sung Huh CPSC 643 Individual Presentation II March 4 th,
Plenoptic Stitching: A Scalable Method for Reconstructing 3D Interactive Walkthroughs Daniel G. Aliaga Ingrid Carlbom
Visual Odometry Michael Adams CS 223B Problem: Measure trajectory of a mobile platform using visual data Mobile Platform (Car) Calibrated Camera.
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Computing motion between images
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
Planar Matchmove Using Invariant Image Features Andrew Kaufman.
CS 376b Introduction to Computer Vision 04 / 01 / 2008 Instructor: Michael Eckmann.
CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from:  David Lowe, UBC  Jiri Matas, CMP Prague
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
Lecture 12: Structure from motion CS6670: Computer Vision Noah Snavely.
Structure Computation. How to compute the position of a point in 3- space given its image in two views and the camera matrices of those two views Use.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
Automatic Camera Calibration
Computer vision: models, learning and inference
Path-Based Constraints for Accurate Scene Reconstruction from Aerial Video Mauricio Hess-Flores 1, Mark A. Duchaineau 2, Kenneth I. Joy 3 Abstract - This.
Sequential Reconstruction Segment-Wise Feature Track and Structure Updating Based on Parallax Paths Mauricio Hess-Flores 1, Mark A. Duchaineau 2, Kenneth.
Satellites in Our Pockets: An Object Positioning System using Smartphones Justin Manweiler, Puneet Jain, Romit Roy Choudhury TsungYun
What Does the Scene Look Like From a Scene Point? Donald Tanguay August 7, 2002 M. Irani, T. Hassner, and P. Anandan ECCV 2002.
WP3 - 3D reprojection Goal: reproject 2D ball positions from both cameras into 3D space Inputs: – 2D ball positions estimated by WP2 – 2D table positions.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
3D SLAM for Omni-directional Camera
Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009]
MESA LAB Multi-view image stitching Guimei Zhang MESA LAB MESA (Mechatronics, Embedded Systems and Automation) LAB School of Engineering, University of.
Visualization of Scene Structure Uncertainty in a Multi-View Reconstruction Pipeline Shawn Recker 1, Mauricio Hess- Flores 1, Mark A. Duchaineau 2, and.
Scientific Writing Abstract Writing. Why ? Most important part of the paper Number of Readers ! Make people read your work. Sell your work. Make your.
Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac.
CSCE 643 Computer Vision: Structure from Motion
Ray Divergence-Based Bundle Adjustment Conditioning for Multi-View Stereo Mauricio Hess-Flores 1, Daniel Knoblauch 2, Mark A. Duchaineau 3, Kenneth I.
© 2005 Martin Bujňák, Martin Bujňák Supervisor : RNDr.
18 th August 2006 International Conference on Pattern Recognition 2006 Epipolar Geometry from Two Correspondences Michal Perďoch, Jiří Matas, Ondřej Chum.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
1 Motion estimation from image and inertial measurements Dennis Strelow and Sanjiv Singh.
Plane-based external camera calibration with accuracy measured by relative deflection angle Chunhui Cui , KingNgiNgan Journal Image Communication Volume.
Feature Matching. Feature Space Outlier Rejection.
3D reconstruction from uncalibrated images
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
MASKS © 2004 Invitation to 3D vision. MASKS © 2004 Invitation to 3D vision Lecture 1 Overview and Introduction.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
Epipolar Geometry and Stereo Vision
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Digital Visual Effects, Spring 2007 Yung-Yu Chuang 2007/4/17
3D Graphics Rendering PPT By Ricardo Veguilla.
Approximate Models for Fast and Accurate Epipolar Geometry Estimation
The Brightness Constraint
Epipolar geometry.
Structure from motion Input: Output: (Tomasi and Kanade)
More on single-view geometry class 10
Session: Video Analysis and Action Recognition, Friday 9 November 2012
Multiple View Geometry for Robotics
Noah Snavely.
Structure from motion Input: Output: (Tomasi and Kanade)
Lecture 15: Structure from motion
Presentation transcript:

Path-Based Constraints for Accurate Scene Reconstruction from Aerial Video Mauricio Hess-Flores1, Mark A. Duchaineau2, Kenneth I. Joy3 1,3Institute for Data Analysis and Visualization, University of California, Davis, USA 2Lawrence Livermore National Laboratory, Livermore, CA, USA** 1mhessf@ucdavis.edu, 2duchaine@google.com, 3kenneth.i.joy@gmail.com **This author is now at Google, Inc. Abstract - This paper discusses the constraints imposed by the path of a moving camera in multi-view sequential scene reconstruction scenarios such as in aerial video, which allow for an efficient detection and correction of inaccuracies in the feature tracking and structure computation processes. The main insight is that for short, planar segments of a continuous camera trajectory, parallax movement corresponding to a viewed scene point should ideally form a scaled and translated version of this trajectory when projected onto a parallel plane. Two constraints arise, which allow for the detection and correction of inaccurate feature tracks and scene structure, which differ from classical approaches such as factorization and RANSAC. Results are shown for real and synthetic aerial video and turntable sequences, where the use of such constraints was proven to correct outlier tracks, detect and correct tracking drift, and allow for a novel improvement of scene structure, while additionally resulting in an improved convergence for bundle adjustment optimization. Introduction Algorithm (continued) Results 1) Bundle adjustment convergence. Total reprojection error ε in pixels, processing time t in seconds and iterations I of Levenberg-Marquardt, for bundle adjustment applied using the output of the proposed algorithm (PPBA) versus bundle adjustment applied using the original feature tracks and structure (TBA), along with number of scene points NSP: Initial parallax path calculation (assuming known cameras) Accurate models developed from aerial video can form a base for large-scale multi-sensor networks that support activities in detection, surveillance, tracking, registration, terrain modeling and ultimately semantic scene analysis. Due to varying lighting conditions, occlusions, repetitive patterns and other issues, feature tracks may not be perfect and this skews subsequent calibration and structure estimation. For short, planar segments of a continuous camera trajectory, parallax movement corresponding to a viewed scene point should ideally form a scaled and translated version of this trajectory, or a parallax path, when projected onto a parallel plane. This introduces two strong constraints, which differ from classical factorization and RANSAC, that can be used to detect and correct inaccurate feature tracks, while allowing for a very simple structure computation. Input images 2) Drift detection and track correction results (Dinosaur dataset): Inaccurate dense reconstruction Position-invariant reference Parallax paths Each path on the reconstruction plane, computed for a given track, is placed in a position-invariant reference, where ideally each differs only by scale: Original tracks Corrected tracks Detected drift 3) Improvement in scene structure (Stockton aerial dataset): Replicas At position-invariant reference, where paths only differ by scale s Original parallax paths Algorithm Inter and intra-camera constraints In this reference, inter-camera consensus path and intra-camera line constraints are defined, whose intersections (perfect grid) predict how inaccurate tracks should be corrected: Position-invariant reference (top), constrained paths (bottom) Path differences from perfect grid Original (top) versus corrected structure (bottom) Algorithm flowchart Ray equation: Future work Track auto-completion, making use of track scale values. Accurate, dense feature tracking despite occlusions. Accurate matching over texture-less regions. Data compression, achieved by storing only scale-related information. Multi-view tensor based on the proposed constraints. Ray-plane intersection: Scaled paths Consensus path Locus line Structure computation after constraint-based track correction C = camera center, P = projection matrix, plane = (A,B,C,D), x = pixel feature track position, X = 3D position This work was supported in part by the Department of Energy, National Nuclear Security Agency through Contract No. DE-GG52-09NA29355. This work was performed in part under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.