Vision Reading Group, 2nd of May 2018 Martin Rünz

Slides:



Advertisements
Similar presentations
Real-Time Template Tracking
Advertisements

DTAM: Dense Tracking and Mapping in Real-Time
Lecture 2: Convolution and edge detection CS4670: Computer Vision Noah Snavely From Sandlot ScienceSandlot Science.
Morphing & Warping 2D Morphing Involves 2 steps 1.Image warping “get features to line up” 2.Cross-dissolve “mix colors” (fade-in/fadeout transition)
Image enhancement in the spatial domain. Human vision for dummies Anatomy and physiology Wavelength Wavelength sensitivity.
|| Dmitry Laptev, Joachim M. Buhmann Machine Learning Lab, ETH Zurich 05/09/14Dmitry Laptev1 Convolutional Decision Trees.
Learning Convolutional Feature Hierarchies for Visual Recognition
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Bayesian Image Super-resolution, Continued Lyndsey C. Pickup, David P. Capel, Stephen J. Roberts and Andrew Zisserman, Robotics Research Group, University.
Unsupervised Learning With Neural Nets Deep Learning and Neural Nets Spring 2015.
Adam Rachmielowski 615 Project: Real-time monocular vision-based SLAM.
Real-time Combined 2D+3D Active Appearance Models Jing Xiao, Simon Baker,Iain Matthew, and Takeo Kanade CVPR 2004 Presented by Pat Chan 23/11/2004.
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
High-Quality Video View Interpolation
1 Image Filtering Readings: Ch 5: 5.4, 5.5, 5.6,5.7.3, 5.8 (This lecture does not follow the book.) Images by Pawan SinhaPawan Sinha formal terminology.
1 Image filtering Images by Pawan SinhaPawan Sinha.
Basic Image Processing January 26, 30 and February 1.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
Xiaoguang Han Department of Computer Science Probation talk – D Human Reconstruction from Sparse Uncalibrated Views.
Effective Computation of Linear Filters The following properties are used to shorten the computation time f  g = g  f, f  (g  h) = (f  g)  h, f 
8-1 Chapter 8: Image Restoration Image enhancement: Overlook degradation processes, deal with images intuitively Image restoration: Known degradation processes;
Peter Henry1, Michael Krainin1, Evan Herbst1,
Geometric Modeling using Polygonal Meshes Lecture 3: Discrete Differential Geometry and its Application to Mesh Processing Office: South B-C Global.
Ch.8 Efficient Coding of Visual Scenes by Grouping and Segmentation Bayesian Brain Tai Sing Lee and Alan L. Yuille Heo, Min-Oh.
Lecture#4 Image reconstruction
Instructor: Mircea Nicolescu Lecture 5 CS 485 / 685 Computer Vision.
CS559: Computer Graphics Lecture 7: Image Warping and Panorama Li Zhang Spring 2008 Most slides borrowed from Yungyu ChuangYungyu Chuang.
Rich feature hierarchies for accurate object detection and semantic segmentation 2014 IEEE Conference on Computer Vision and Pattern Recognition Ross Girshick,
Deep Learning Overview Sources: workshop-tutorial-final.pdf
CSC2535: Lecture 4: Autoencoders, Free energy, and Minimum Description Length Geoffrey Hinton.
MAN-522 Computer Vision Spring
Miguel Tavares Coimbra
- photometric aspects of image formation gray level images
Summary of “Efficient Deep Learning for Stereo Matching”
Estimating Parametric and Layered Motion
Motion and Optical Flow
ECE 692 – Advanced Topics in Computer Vision
Registration of Pathological Images
Computational Photography Derek Hoiem, University of Illinois
The Brightness Constraint
Fei Li Jinjun Xiong University of Wisconsin-Madison
EECS 274 Computer Vision Stereopsis.
- photometric aspects of image formation gray level images
Lecture 2: Edge detection
Outline Linear Shift-invariant system Linear filters
The Brightness Constraint
Image filtering Hybrid Images, Oliva et al.,
Jeremy Bolton, PhD Assistant Teaching Professor
CNNs and compressive sensing Theoretical analysis
PixelGAN Autoencoders
SBNet: Sparse Blocks Network for Fast Inference
The Brightness Constraint
Towards Understanding the Invertibility of Convolutional Neural Networks Anna C. Gilbert1, Yi Zhang1, Kibok Lee1, Yuting Zhang1, Honglak Lee1,2 1University.
Goodfellow: Chapter 14 Autoencoders
Basic Image Processing
Image filtering
Image filtering
Coupled Horn-Schunck and Lukas-Kanade for image processing
Edge Detection in Computer Vision
Autoencoders Supervised learning uses explicit labels/correct output in order to train a network. E.g., classification of images. Unsupervised learning.
Image Filtering Readings: Ch 5: 5. 4, 5. 5, 5. 6, , 5
Report Yang Zhang.
Scalable light field coding using weighted binary images
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
Occlusion and smoothness probabilities in 3D cluttered scenes
Sign Language Recognition With Unsupervised Feature Learning
Image Morphing using mesh warp and feature based warp
Directional Occlusion with Neural Network
Goodfellow: Chapter 14 Autoencoders
Presentation transcript:

Vision Reading Group, 2nd of May 2018 Martin Rünz CodeSLAM — Learning a Compact, Optimisable Representation for Dense Visual SLAM Vision Reading Group, 2nd of May 2018 Martin Rünz

An alternative representation to depth maps is presented In a nutshell An alternative representation to depth maps is presented

Background – Common Map Representations Surfels Mesh Pose-Graph + Depth-Maps TSDF

Ideas Depth-maps are not random

Ideas Depth-maps are not random (especially in man-made environments)

Ideas Depth-maps are not random A compact representation – code – can be learned by encoder Decoders are differentiable → code can be optimised (w.r.t. photometric loss...) This is useful in SfM or SLAM scenarios Reconstruct high-frequency details from color

Depth Reconstruction Depth-from-mono Depth-from-mono + code Given this differentiable function, warping constraints can be used to optimise c and pose

Architecture Laplace distribution Color feature extractor + uncertainty predictor Laplace distribution Only for training, ground-truth depth Decoder: CODE Variational autoencoder, to increase smoothness of mapping between code and depth Small changes in code → small changes in depth

Video

Inverse depth parametrization average Error behaves more Gaussian original depth

Warping Both functions differentiable to inputs Transform → B 3D Photometric error: Both functions differentiable to inputs Expensive (convolutions) Pre-computed if decoder is linear!

Application SfM Cost function: Initialise poses and code with zero-vector Use residuals + Jacobians in Gauss-Newton style optimisation Cost function: Functionality of : Mask invalid correspondences Relative weighting Huber weighting Down-weight slanted surfaces Down-weight occluded pixels

Experiments SfM

Experiments SfM

Experiments SLAM

Experiments Setups

Experiments Setups

Experiments Influence code entries

Video

Thanks for listening!