INDOOR DENSE DEPTH MAP AT DRONE HOVERING

Slides:



Advertisements
Similar presentations
Registration for Robotics Kurt Konolige Willow Garage Stanford University Patrick Mihelich JD Chen James Bowman Helen Oleynikova Freiburg TORO group: Giorgio.
Advertisements

Efficient High-Resolution Stereo Matching using Local Plane Sweeps Sudipta N. Sinha, Daniel Scharstein, Richard CVPR 2014 Yongho Shin.
L1 sparse reconstruction of sharp point set surfaces
Verification of specifications and aptitude for short-range applications of the Kinect v2 depth sensor Cecilia Chen, Cornell University Lewis’ Educational.
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
KinectFusion: Real-Time Dense Surface Mapping and Tracking
Joint Optimisation for Object Class Segmentation and Dense Stereo Reconstruction Ľubor Ladický, Paul Sturgess, Christopher Russell, Sunando Sengupta, Yalin.
A Keystone-free Hand-held Mobile Projection System Li Zhaorong And KH Wong Reference: Zhaorong Li, Kin-Hong Wong, Yibo Gong, and Ming-Yuen Chang, “An Effective.
Hilal Tayara ADVANCED INTELLIGENT ROBOTICS 1 Depth Camera Based Indoor Mobile Robot Localization and Navigation.
Gratuitous Picture US Naval Artillery Rangefinder from World War I (1918)!!
Segmentation of Floor in Corridor Images for Mobile Robot Navigation Yinxiao Li Clemson University Committee Members: Dr. Stanley Birchfield (Chair) Dr.
Extracting Minimalistic Corridor Geometry from Low-Resolution Images Yinxiao Li, Vidya, N. Murali, and Stanley T. Birchfield Department of Electrical and.
Patch to the Future: Unsupervised Visual Prediction
Optimization & Learning for Registration of Moving Dynamic Textures Junzhou Huang 1, Xiaolei Huang 2, Dimitris Metaxas 1 Rutgers University 1, Lehigh University.
Robust Higher Order Potentials For Enforcing Label Consistency
CVR05 University of California Berkeley 1 Familiar Configuration Enables Figure/Ground Assignment in Natural Scenes Xiaofeng Ren, Charless Fowlkes, Jitendra.
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
CS664 Lecture #19: Layers, RANSAC, panoramas, epipolar geometry Some material taken from:  David Lowe, UBC  Jiri Matas, CMP Prague
The Terrapins Computer Vision Laboratory University of Maryland.
Accurate, Dense and Robust Multi-View Stereopsis Yasutaka Furukawa and Jean Ponce Presented by Rahul Garg and Ryan Kaminsky.
Matthew Brown University of British Columbia (prev.) Microsoft Research [ Collaborators: † Simon Winder, *Gang Hua, † Rick Szeliski † =MS Research, *=MS.
Projective Texture Atlas for 3D Photography Jonas Sossai Júnior Luiz Velho IMPA.
Summer School on Image Processing 2009, Debrecen, Hungary Colour image processing for SHADOW REMOVAL Alina Elena Oprea, University Politehnica of Bucharest.
Online Tracking of Outdoor Lighting Variations for Augmented Reality with Moving Cameras Yanli Liu 1,2 and Xavier Granier 2,3,4 1: College of Computer.
Automatic Registration of Color Images to 3D Geometry Computer Graphics International 2009 Yunzhen Li and Kok-Lim Low School of Computing National University.
Xiaoguang Han Department of Computer Science Probation talk – D Human Reconstruction from Sparse Uncalibrated Views.
3D SLAM for Omni-directional Camera
Presented by: 15 Oct 2001 Neptec Laser Camera System Preliminary Post Flight Results Neptec Design Group.
ALIGNMENT OF 3D ARTICULATE SHAPES. Articulated registration Input: Two or more 3d point clouds (possibly with connectivity information) of an articulated.
The Collocation of Measurement Points in Large Open Indoor Environment Kaikai Sheng, Zhicheng Gu, Xueyu Mao Xiaohua Tian, Weijie Wu, Xiaoying Gan Department.
I 3D: Interactive Planar Reconstruction of Objects and Scenes Adarsh KowdleYao-Jen Chang Tsuhan Chen School of Electrical and Computer Engineering Cornell.
ECE738 Advanced Image Processing Face Detection IEEE Trans. PAMI, July 1997.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
The 18th Meeting on Image Recognition and Understanding 2015/7/29 Depth Image Enhancement Using Local Tangent Plane Approximations Kiyoshi MatsuoYoshimitsu.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
Peter Henry1, Michael Krainin1, Evan Herbst1,
Removing Moving Objects from Point Cloud Scenes Krystof Litomisky and Bir Bhanu International Workshop on Depth Image Analysis November 11, 2012.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
Image-Based Segmentation of Indoor Corridor Floors for a Mobile Robot
Joint Tracking of Features and Edges STAN BIRCHFIELD AND SHRINIVAS PUNDLIK CLEMSON UNIVERSITY ABSTRACT LUCAS-KANADE AND HORN-SCHUNCK JOINT TRACKING OF.
Pixel Clustering and Hyperspectral Image Segmentation for Ocean Colour Remote Sensing Xuexing Zeng 1, Jinchang Ren 1, David Mckee 2 Samantha Lavender 3.
Yizhou Yu Texture-Mapping Real Scenes from Photographs Yizhou Yu Computer Science Division University of California at Berkeley Yizhou Yu Computer Science.
High Resolution Surface Reconstruction from Overlapping Multiple-Views
Line Matching Jonghee Park GIST CV-Lab..  Lines –Fundamental feature in many computer vision fields 3D reconstruction, SLAM, motion estimation –Useful.
Real-Time Feature Matching using Adaptive and Spatially Distributed Classification Trees Aurélien BOFFY, Yanghai TSIN, Yakup GENC.
Photoconsistency constraint C2 q C1 p l = 2 l = 3 Depth labels If this 3D point is visible in both cameras, pixels p and q should have similar intensities.
ICCV 2007 Optimization & Learning for Registration of Moving Dynamic Textures Junzhou Huang 1, Xiaolei Huang 2, Dimitris Metaxas 1 Rutgers University 1,
Presented by 翁丞世  View Interpolation  Layered Depth Images  Light Fields and Lumigraphs  Environment Mattes  Video-Based.
Modeling Perspective Effects in Photographic Composition Zihan Zhou, Siqiong He, Jia Li, and James Z. Wang The Pennsylvania State University.
By: Rachel Yuen, Chad Van De Hey, and Jake Trotman
Signal and Image Processing Lab
Karel Lebeda, Simon Hadfield, Richard Bowden
Real-Time Soft Shadows with Adaptive Light Source Sampling
Range Image Segmentation for Modeling and Object Detection in Urban Scenes Cecilia Chen & Ioannis Stamos Computer Science Department Graduate Center, Hunter.
Capturing, Processing and Experiencing Indian Monuments
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
Reconstruction For Rendering distribution Effect
Compositional Human Pose Regression
Saliency detection with background model
Fast Preprocessing for Robust Face Sketch Synthesis
SemanticFusion: Dense 3D Semantic Mapping with Convolutional Neural Networks Paper by John McCormac, Ankur Handa, Andrew Davison, and Stefan Leutenegger.
CVPR 2014 Orientational Pyramid Matching for Recognizing Indoor Scenes
Paper Presentation Aryeh Zapinsky
Combining Geometric- and View-Based Approaches for Articulated Pose Estimation David Demirdjian MIT Computer Science and Artificial Intelligence Laboratory.
Online Graph-Based Tracking
Topological Signatures For Fast Mobility Analysis
Learning complex visual concepts
Artistic Rendering Final Project Initial Proposal
Presentation transcript:

INDOOR DENSE DEPTH MAP AT DRONE HOVERING Arindam Saha, Soumyadip Maity, Brojeshwar Bhowmick 8th October 2018

Content Motivation & Problem Statement Structure from Small Motion (SfSM) Literature survey Proposed Methodology Novelty Detail description (Segmentation & Depth Propagation) Experimental Result

Motivation & Problem Statement Drone Capture Near Obstacle vSLAM produces sparse point cloud Open space Near Far Real-time dense reconstruction or dense visual SLAM is computationally very heavy. So there exist a requirement of a computationally light weight real-time solution for dense depth estimation. But this is actually require

Structure from Small Motion (SfSM) Dense depth map estimation on SfSM configuration is computationally less complex. So dense depth estimation at drone hovering using SfSM configuration (small baseline images) is a feasible approach.

Limitations in State of the Art (SoA) Traditional visual SLAMs [ORB, cvpr 14; PL, ICRA 17; Edge, ICCV 2017] produce erroneous metric depth when the baseline is less than ∼8 mm and viewing angle of a 3D point is less than 0.2◦ [Yu F., cvpr 14]. There exists depth estimation inaccuracy in SfSM system on planer surfaces, object boundary etc. Execution time more than several minutes.

Limitation in SoA (Cont.) Near Far Reference Image (Lab-room dataset) Kinect Depth Map Depth Map using S. Im et al. ICCV 15 Depth Map using H. Ha et al. CVPR 16

Proposed Methodology Sparse reconstruction using Edge SLAM while drone is on move. Outputs are sparse point cloud and key frame poses. Estimate the poses (without BA) of small motion images (at drone hovering) using Edge SLAM. Segment the visible space with sudden depth changes. Dense depth propagation for every segment independently.

Novelty The visible space is segmented based on sparse reconstruction and depth propagation runs independently on every segment. Patch based plane fitting approach is proposed through minimizing a cost function consisting of sparse 3D points, intensity and gradients of pixels and co-planarity with neighbouring patches. We attach a weight with cost of every patch based on its depth initialization. Further add that cost for minimization when the weight is below to certain threshold.

Segmentation Segment the visible 3D space based on sparse point cloud using a combination of nearest neighbour clustering and colour similarity between 3D points. Segment the reference image using a colour based region growing segmentation. Reference Image

Depth Propagation Cost Function = Geometric Consistency Gradient Matching Pairwise Regularization Photo Consistency ω represents the set patches in a segment. Nω is the set neighbouring patches of patch p. φp represents the set of sparse 3D points whose projection lies within the patch p boundary and η = |φp | and area(φp) represents the 2D projected area of points in φp and ∆ is a given threshold.

Depth Propagation (Cont.) Geometric consistency = ∂ denotes the average reprojection error for the point i ∈ φp. It provides lower weight to the sparse 3D points having high reprojection error. Photo-consistency = θ is the set of pixels in patch p.

Depth Propagation (Cont.) (b) (c) (d) Regularization =

Results We perform all our experiments on 640x480 VGA resolution. Our method takes ∼14 sec using 20 images. Near Far Our depth estimation in Lab-room dataset

Results (Cont.) Office-Space dataset Kinect S. Im et al. ICCV 15 H. Ha et al. CVPR 16 Ours The mean error in depth estimation using our method against Kinect point cloud is 0.2352 meter whereas method by H. Ha et al. CVPR 16 producing 2.11575 meters errors.

Results (Cont.) Depth map using our method Objects are identified successfully S. Im et al. ICCV 15 H. Ha et al. CVPR 16

Thank You