Path-Based Constraints for Accurate Scene Reconstruction from Aerial Video Mauricio Hess-Flores 1, Mark A. Duchaineau 2, Kenneth I. Joy 3 Abstract - This.

Slides:



Advertisements
Similar presentations
Vanishing points  .
Advertisements

A Robust Super Resolution Method for Images of 3D Scenes Pablo L. Sala Department of Computer Science University of Toronto.
The fundamental matrix F
CSE473/573 – Stereo and Multiple View Geometry
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam Road, Hong Kong Three-dimensional curve reconstruction from.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Chapter 6 Feature-based alignment Advanced Computer Vision.
Tracking Multiple Occluding People by Localizing on Multiple Scene Planes Saad M. Khan and Mubarak Shah, PAMI, VOL. 31, NO. 3, MARCH 2009, Donguk Seo
Active Calibration of Cameras: Theory and Implementation Anup Basu Sung Huh CPSC 643 Individual Presentation II March 4 th,
Last Time Pinhole camera model, projection
Plenoptic Stitching: A Scalable Method for Reconstructing 3D Interactive Walkthroughs Daniel G. Aliaga Ingrid Carlbom
Visual Odometry Michael Adams CS 223B Problem: Measure trajectory of a mobile platform using visual data Mobile Platform (Car) Calibrated Camera.
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Computing motion between images
Flexible Bump Map Capture From Video James A. Paterson and Andrew W. Fitzgibbon University of Oxford Calibration Requirement:
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
Planar Matchmove Using Invariant Image Features Andrew Kaufman.
CS 376b Introduction to Computer Vision 04 / 01 / 2008 Instructor: Michael Eckmann.
CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from:  David Lowe, UBC  Jiri Matas, CMP Prague
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Lecture 12: Structure from motion CS6670: Computer Vision Noah Snavely.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
Automatic Camera Calibration
Computer vision: models, learning and inference
CSC 589 Lecture 22 Image Alignment and least square methods Bei Xiao American University April 13.
Sequential Reconstruction Segment-Wise Feature Track and Structure Updating Based on Parallax Paths Mauricio Hess-Flores 1, Mark A. Duchaineau 2, Kenneth.
What Does the Scene Look Like From a Scene Point? Donald Tanguay August 7, 2002 M. Irani, T. Hassner, and P. Anandan ECCV 2002.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
WP3 - 3D reprojection Goal: reproject 2D ball positions from both cameras into 3D space Inputs: – 2D ball positions estimated by WP2 – 2D table positions.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
A Local Adaptive Approach for Dense Stereo Matching in Architectural Scene Reconstruction C. Stentoumis 1, L. Grammatikopoulos 2, I. Kalisperakis 2, E.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
3D SLAM for Omni-directional Camera
Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009]
MESA LAB Multi-view image stitching Guimei Zhang MESA LAB MESA (Mechatronics, Embedded Systems and Automation) LAB School of Engineering, University of.
The Correspondence Problem and “Interest Point” Detection Václav Hlaváč Center for Machine Perception Czech Technical University Prague
Visualization of Scene Structure Uncertainty in a Multi-View Reconstruction Pipeline Shawn Recker 1, Mauricio Hess- Flores 1, Mark A. Duchaineau 2, and.
Scientific Writing Abstract Writing. Why ? Most important part of the paper Number of Readers ! Make people read your work. Sell your work. Make your.
An Information Fusion Approach for Multiview Feature Tracking Esra Ataer-Cansizoglu and Margrit Betke ) Image and.
Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac.
Visibility map (high low) Surveillance with Visual Tagging and Camera Placement J. Zhao and S.-C. Cheung — Center for Visualization and Virtual Environment,
Ray Divergence-Based Bundle Adjustment Conditioning for Multi-View Stereo Mauricio Hess-Flores 1, Daniel Knoblauch 2, Mark A. Duchaineau 3, Kenneth I.
© 2005 Martin Bujňák, Martin Bujňák Supervisor : RNDr.
18 th August 2006 International Conference on Pattern Recognition 2006 Epipolar Geometry from Two Correspondences Michal Perďoch, Jiří Matas, Ondřej Chum.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
Plane-based external camera calibration with accuracy measured by relative deflection angle Chunhui Cui , KingNgiNgan Journal Image Communication Volume.
Feature Matching. Feature Space Outlier Rejection.
3D reconstruction from uncalibrated images
Computer vision: models, learning and inference M Ahad Multiple Cameras
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
Last Two Lectures Panoramic Image Stitching
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
MASKS © 2004 Invitation to 3D vision. MASKS © 2004 Invitation to 3D vision Lecture 1 Overview and Introduction.
Lecture 22: Structure from motion CS6670: Computer Vision Noah Snavely.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
Epipolar Geometry and Stereo Vision
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Mauricio Hess-Flores1, Mark A. Duchaineau2, Kenneth I. Joy3
Epipolar geometry.
Structure from motion Input: Output: (Tomasi and Kanade)
Persistent Surveillance
Session: Video Analysis and Action Recognition, Friday 9 November 2012
Structure from motion Input: Output: (Tomasi and Kanade)
Presentation transcript:

Path-Based Constraints for Accurate Scene Reconstruction from Aerial Video Mauricio Hess-Flores 1, Mark A. Duchaineau 2, Kenneth I. Joy 3 Abstract - This paper discusses the constraints imposed by the path of a moving camera in multi-view sequential scene reconstruction scenarios such as in aerial video, which allow for an efficient detection and correction of inaccuracies in the feature tracking and structure computation processes. The main insight is that for short, planar segments of a continuous camera trajectory, parallax movement corresponding to a viewed scene point should ideally form a scaled and translated version of this trajectory when projected onto a parallel plane. Two constraints arise, which allow for the detection and correction of inaccurate feature tracks and scene structure, which differ from classical approaches such as factorization and RANSAC. Results are shown for real and synthetic aerial video and turntable sequences, where the use of such constraints was proven to correct outlier tracks, detect and correct tracking drift, and allow for a novel improvement of scene structure, while additionally resulting in an improved convergence for bundle adjustment optimization. 1,3 Institute for Data Analysis and Visualization, University of California, Davis, USA Introduction Algorithm (continued) Results Future work Accurate models developed from aerial video can form a base for large-scale multi-sensor networks that support activities in detection, surveillance, tracking, registration, terrain modeling and ultimately semantic scene analysis. Due to varying lighting conditions, occlusions, repetitive patterns and other issues, feature tracks may not be perfect and this skews subsequent calibration and structure estimation. For short, planar segments of a continuous camera trajectory, parallax movement corresponding to a viewed scene point should ideally form a scaled and translated version of this trajectory, or a parallax path, when projected onto a parallel plane. This introduces two strong constraints, which differ from classical factorization and RANSAC, that can be used to detect and correct inaccurate feature tracks, while allowing for a very simple structure computation Track auto-completion, making use of track scale values. Accurate, dense feature tracking despite occlusions. Accurate matching over texture-less regions. Data compression, achieved by storing only scale-related information. Multi-view tensor based on the proposed constraints. This work was supported in part by the Department of Energy, National Nuclear Security Agency through Contract No. DE-GG52-09NA This work was performed in part under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA Lawrence Livermore National Laboratory, Livermore, CA, USA** Each path on the reconstruction plane, computed for a given track, is placed in a position-invariant reference, where ideally each differs only by scale: 1) Bundle adjustment convergence. Total reprojection error ε in pixels, processing time t in seconds and iterations I of Levenberg-Marquardt, for bundle adjustment applied using the output of the proposed algorithm (PPBA) versus bundle adjustment applied using the original feature tracks and structure (TBA), along with number of scene points N SP : Input images Algorithm flowchart Consensus path Locus line Scaled paths 2) Drift detection and track correction results (Dinosaur dataset): Ray equation: C = camera center, P = projection matrix, plane = (A,B,C,D), x = pixel feature track position, X = 3D position Position-invariant reference Algorithm **This author is now at Google, Inc. Inaccurate dense reconstruction Parallax paths Ray-plane intersection: Initial parallax path calculation (assuming known cameras) Original parallax paths At position-invariant reference, where paths only differ by scale s Inter and intra-camera constraints In this reference, inter-camera consensus path and intra-camera line constraints are defined, whose intersections (perfect grid) predict how inaccurate tracks should be corrected: Replicas Structure computation after constraint-based track correction Original tracks Corrected tracksDetected drift 3) Improvement in scene structure (Stockton aerial dataset): Position-invariant reference (top), constrained paths (bottom) Path differences from perfect grid Original (top) versus corrected structure (bottom)