For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.

Slides:



Advertisements
Similar presentations
Miroslav Hlaváč Martin Kozák Fish position determination in 3D space by stereo vision.
Advertisements

The fundamental matrix F
CSE473/573 – Stereo and Multiple View Geometry
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Recent work in image-based rendering from unstructured image collections and remaining challenges Sudipta N. Sinha Microsoft Research, Redmond, USA.
Computer vision: models, learning and inference
Parallel Tracking and Mapping for Small AR Workspaces Vision Seminar
A Global Linear Method for Camera Pose Registration
Discrete-Continuous Optimization for Large-scale Structure from Motion David Crandall, Andrew Owens, Noah Snavely, Dan Huttenlocher Presented by: Rahul.
Structure from motion.
Constructing immersive virtual space for HAI with photos Shingo Mori Yoshimasa Ohmoto Toyoaki Nishida Graduate School of Informatics Kyoto University GrC2011.
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Last Time Pinhole camera model, projection
Multi video camera calibration and synchronization.
Adam Rachmielowski 615 Project: Real-time monocular vision-based SLAM.
Contents Description of the big picture Theoretical background on this work The Algorithm Examples.
Postcalibrating RBLFs Vaibhav Vaish. A “Really Big Light Field” 1300x1030 color images 62x56 viewpoints per slab Seven slabs of 3472 images each
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Multi-view stereo Many slides adapted from S. Seitz.
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
The plan for today Camera matrix
CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from:  David Lowe, UBC  Jiri Matas, CMP Prague
Constructing immersive virtual space for HAI with photos Shingo Mori Yoshimasa Ohmoto Toyoaki Nishida Graduate School of Informatics Kyoto University GrC2011.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
Lecture 12: Structure from motion CS6670: Computer Vision Noah Snavely.
Accurate, Dense and Robust Multi-View Stereopsis Yasutaka Furukawa and Jean Ponce Presented by Rahul Garg and Ryan Kaminsky.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Automatic Camera Calibration
Path-Based Constraints for Accurate Scene Reconstruction from Aerial Video Mauricio Hess-Flores 1, Mark A. Duchaineau 2, Kenneth I. Joy 3 Abstract - This.
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
Last Week Recognized the fact that the 2D image is a representation of a 3D scene thus contains a consistent interpretation –Labeled edges –Labeled vertices.
Automatic Registration of Color Images to 3D Geometry Computer Graphics International 2009 Yunzhen Li and Kok-Lim Low School of Computing National University.
KinectFusion : Real-Time Dense Surface Mapping and Tracking IEEE International Symposium on Mixed and Augmented Reality 2011 Science and Technology Proceedings.
Flow Separation for Fast and Robust Stereo Odometry [ICRA 2009]
CSCE 643 Computer Vision: Structure from Motion
3D Reconstruction Jeff Boody. Goals ● Reconstruct 3D models from a sequence of at least two images ● No prior knowledge of the camera or scene ● Use the.
Ray Divergence-Based Bundle Adjustment Conditioning for Multi-View Stereo Mauricio Hess-Flores 1, Daniel Knoblauch 2, Mark A. Duchaineau 3, Kenneth I.
Multi-View Stereo : A Parametric Study Number of images, their angular separation and the like.
Vision-based human motion analysis: An overview Computer Vision and Image Understanding(2007)
Acquiring 3D models of objects via a robotic stereo head David Virasinghe Department of Computer Science University of Adelaide Supervisors: Mike Brooks.
Asian Institute of Technology
© 2005 Martin Bujňák, Martin Bujňák Supervisor : RNDr.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Senior Project Poster Day 2005, CIS Dept. University of Pennsylvania Surface Reconstruction from Feature Based Stereo Mickey Mouse, Donald Duck Faculty.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
1 Motion estimation from image and inertial measurements Dennis Strelow and Sanjiv Singh.
Feature Matching. Feature Space Outlier Rejection.
776 Computer Vision Jan-Michael Frahm & Enrique Dunn Spring 2013.
Fast Semi-Direct Monocular Visual Odometry
Visual Odometry David Nister, CVPR 2004
High Resolution Surface Reconstruction from Overlapping Multiple-Views
Presented by: Idan Aharoni
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
1 Long-term image-based motion estimation Dennis Strelow and Sanjiv Singh.
MASKS © 2004 Invitation to 3D vision. MASKS © 2004 Invitation to 3D vision Lecture 1 Overview and Introduction.
Paper – Stephen Se, David Lowe, Jim Little
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Real Time Dense 3D Reconstructions: KinectFusion (2011) and Fusion4D (2016) Eleanor Tursman.
Mauricio Hess-Flores1, Mark A. Duchaineau2, Kenneth I. Joy3
Zhigang Zhu, K. Deepak Rajasekar Allen R. Hanson, Edward M. Riseman
Modeling the world with photos
Structure from motion Input: Output: (Tomasi and Kanade)
Multiple View Geometry for Robotics
Image processing and computer vision
Noah Snavely.
Introduction Computer vision is the analysis of digital images
Structure from motion Input: Output: (Tomasi and Kanade)
INDOOR DENSE DEPTH MAP AT DRONE HOVERING
Presentation transcript:

For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam CTT IN, Bangalore

For Internal Use Only. © CT T IN EM. All rights reserved. Project Goal Volume Estimation of mine dumps Infrastructure development monitoring Augmented Reality Page 2 3D capture of ground structures using aerial imagery

For Internal Use Only. © CT T IN EM. All rights reserved. 3D from Images : Stereo? Page 3

For Internal Use Only. © CT T IN EM. All rights reserved. Stereo 3D information can be ascertained if an object is visible from two views separated by a baseline This helps us to estimate the depth of the scene Page 4

For Internal Use Only. © CT T IN EM. All rights reserved. Disparity/ Depth Image Page 5 Stereo Input Images Disparity / Depth Image

For Internal Use Only. © CT T IN EM. All rights reserved. Multi View Stereo (MVS) Images from multiple views at short baselines used. Give Better Precision and reduce matching ambiguity Case for Multi View Stereo Disparity baseline, focal length and matching. Page 6 Camera Model Needed !

For Internal Use Only. © CT T IN EM. All rights reserved. Calibration of a Camera Model Internal parameters Focal length, pixel aspect ratio etc External camera parameters Rotation and Translation in global frame of reference Page 7 Calibration: finding the internal parameters of the camera

For Internal Use Only. © CT T IN EM. All rights reserved. STRUCTURE FROM MOTION Page 8

For Internal Use Only. © CT T IN EM. All rights reserved. Structure from Motion (SFM) Finding the complete 3D object model and complete camera parameters from a collection of images taken from various view- points. Involves  Stereo Initialization  Triangulation  Bundle Adjustment. Page 9

For Internal Use Only. © CT T IN EM. All rights reserved. Bundle Adjustment Stereo Initialization: Finding relation between features in two initial scenes. Bundle Adjustment: Iteratively minimizing reprojection error while adding more cameras and views. Page 10 Computationally Expensive ! Initialization is Key

For Internal Use Only. © CT T IN EM. All rights reserved. SFM: Reconstruction Page 11 SFM: 2 imagesSFM: 5 imagesSFM: 20 images Clearly, not suitable for dense reconstruction.

For Internal Use Only. © CT T IN EM. All rights reserved. SFM -> Multi-View Stereo Pipeline SFM Typically involves matching of sparse features and triangulation of those features. Generates Camera Parameters. Multi-View Stereo Patch based “every pixel” methods used to estimate the disparity/ depth for the whole of a scene. Uses Camera Parameters to give dense depth estimates. Page 12 SFM to MVS pipeline gives dense reconstructions !

For Internal Use Only. © CT T IN EM. All rights reserved. Accurate, Dense and Robust MVS  Extract features  Get a sparse set of initial matches  Iteratively expand matches to nearby locations  Use visibility constraints to filter out false matches Page 13

For Internal Use Only. © CT T IN EM. All rights reserved. The Missing Link Page 14 Multi View Stereo SFM Images Where do the Images come from ?

For Internal Use Only. © CT T IN EM. All rights reserved. LOCALIZING THE CAMERA Page 15

For Internal Use Only. © CT T IN EM. All rights reserved. PTAM: Parallel Tracking and Mapping Page 16 Stereo Initialization Tracking Mapping PTAM: Key frame selection

For Internal Use Only. © CT T IN EM. All rights reserved. PTAM Tracking and mapping are done in parallel allowing more features to be added to map as they are detected. Bundle Adjustment is done after every few frames. Enforces a pose change and time heuristic to select key frames. Page 17

For Internal Use Only. © CT T IN EM. All rights reserved. KeyFrames Page 18

For Internal Use Only. © CT T IN EM. All rights reserved. PTAM – Pose Page 19

For Internal Use Only. © CT T IN EM. All rights reserved. PTAM -> SFM -> MVS Block Results Page 20 CUP_60 dataset

For Internal Use Only. © CT T IN EM. All rights reserved. PTAM -> SFM -> MVS Block Results Page 21 Olympic Coke CAN

For Internal Use Only. © CT T IN EM. All rights reserved. PTAM -> SFM -> MVS Block Results Page 22 Olympic Coke CAN + Pen

For Internal Use Only. © CT T IN EM. All rights reserved. System Block Diagram – So Far Page 23 Multi View Stereo SFM Keyframes PTAM Bundler PMVS-2 3 stage dense reconstruction pipeline

For Internal Use Only. © CT T IN EM. All rights reserved. Volume Estimation 3D reconstructions stored as point clouds, a set of points in space with color information. From a point cloud, planar features are segmented out. Remaining points are clustered. User views clusters and gives the reference ground truth data and the cluster whose volume is to be estimated. Page 24

For Internal Use Only. © CT T IN EM. All rights reserved. Segmentation and Filtering Page 25

For Internal Use Only. © CT T IN EM. All rights reserved. Volume Estimation After segmenting the point cloud, the volume is estimated by finding the convex hull of the 3-D point cloud. Page 26

For Internal Use Only. © CT T IN EM. All rights reserved. Volume Estimation Page 27 Original Point cloud Clusters

For Internal Use Only. © CT T IN EM. All rights reserved. Volume Estimation - Dataset Ground Truth data : 16.2 cm distance between pens Height of Cylinder : 12.9 cm Radius of Cylinder : 2.9 cm Volume of Cylinder : Page 28

For Internal Use Only. © CT T IN EM. All rights reserved. Volume Estimation - Dataset Volume for PTAM dataset: cu cm Image Resolution: 640 x 480 Accuracy : ground truth is 85.4 % of volume Number of Images: 102 Volume for DSLR dataset: cu cm Image Resolution: 1920x1480 Accuracy : ground truth is 81.4 % of volume Number of Images: 30 Page 29

For Internal Use Only. © CT T IN EM. All rights reserved. Volume Accuracy The multi view stereo algorithm gives 98.7% of points 1.25 mm of the reconstruction for reference datasets. Cameras parameters are noisy, affecting volume accuracy. Pose information given by the IMU can improve camera parameters. Clustering done without a-priori shape information, if given, outliers can be filtered out and geometric consistency enforced. Page 30

For Internal Use Only. © CT T IN EM. All rights reserved. Scope for Improvement 1.Use sensor data from IMU to estimate camera pose 2. Make it a real time, live dense reconstruction system 3. Improve accuracy of volume estimation 4. Plan the flight of the UAV doing the reconstruction 5.Making the reconstruction interactive Page 31

For Internal Use Only. © CT T IN EM. All rights reserved. Related work Dense Reconstruction on the fly (TU Graz) :  Real time reconstruction  User interaction with live reconstruction  Successfully adapted to UAV Dense Tracking and Mapping (Imperial College, UK):  Real time dense reconstruction using GPU  Superior Tracking performance, blur resistant Live dense reconstruction from Monocular Camera (IC) :  Real time monocular dense reconstruction  Sparse Tracking Page 32

For Internal Use Only. © CT T IN EM. All rights reserved. THANK YOU ! Page 33