Toward Object Discovery and Modeling via 3-D Scene Comparison Evan Herbst, Peter Henry, Xiaofeng Ren, Dieter Fox University of Washington; Intel Research.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Probabilistic Tracking and Recognition of Non-rigid Hand Motion
Caroline Rougier, Jean Meunier, Alain St-Arnaud, and Jacqueline Rousseau IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 21, NO. 5,
Analysis of Contour Motions Ce Liu William T. Freeman Edward H. Adelson Computer Science and Artificial Intelligence Laboratory Massachusetts Institute.
Dynamic Bayesian Networks (DBNs)
Patch to the Future: Unsupervised Visual Prediction
Qualifying Exam: Contour Grouping Vida Movahedi Supervisor: James Elder Supervisory Committee: Minas Spetsakis, Jeff Edmonds York University Summer 2009.
GrabCut Interactive Image (and Stereo) Segmentation Carsten Rother Vladimir Kolmogorov Andrew Blake Antonio Criminisi Geoffrey Cross [based on Siggraph.
Proximity Computations between Noisy Point Clouds using Robust Classification 1 Jia Pan, 2 Sachin Chitta, 1 Dinesh Manocha 1 UNC Chapel Hill 2 Willow Garage.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Localization David Johnson cs6370. Basic Problem Go from thisto this.
Xiaofeng Ren: Intel Labs Seattle Dieter Fox: UW and Intel Labs Seattle Kurt Konolige: Willow Garage Jana Kosecka: George Mason University.
Paper Discussion: “Simultaneous Localization and Environmental Mapping with a Sensor Network”, Marinakis et. al. ICRA 2011.
Tracking Objects with Dynamics Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 04/21/15 some slides from Amin Sadeghi, Lana Lazebnik,
A Data-Driven Approach to Quantifying Natural Human Motion SIGGRAPH ’ 05 Liu Ren, Alton Patrick, Alexei A. Efros, Jassica K. Hodgins, and James M. Rehg.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Region-Level Motion- Based Background Modeling and Subtraction Using MRFs Shih-Shinh Huang Li-Chen Fu Pei-Yung Hsiao 2007 IEEE.
High-Quality Video View Interpolation
Particle Filters for Mobile Robot Localization 11/24/2006 Aliakbar Gorji Roborics Instructor: Dr. Shiri Amirkabir University of Technology.
CSE 291 Final Project: Adaptive Multi-Spectral Differencing Andrew Cosand UCSD CVRR.
MULTIPLE MOVING OBJECTS TRACKING FOR VIDEO SURVEILLANCE SYSTEMS.
High Speed Obstacle Avoidance using Monocular Vision and Reinforcement Learning Jeff Michels Ashutosh Saxena Andrew Y. Ng Stanford University ICML 2005.
Goal: Fast and Robust Velocity Estimation P1P1 P2P2 P3P3 P4P4 Our Approach: Alignment Probability ●Spatial Distance ●Color Distance (if available) ●Probability.
Object Detection Using the Statistics of Parts Henry Schneiderman Takeo Kanade Presented by : Sameer Shirdhonkar December 11, 2003.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Handwritten Character Recognition using Hidden Markov Models Quantifying the marginal benefit of exploiting correlations between adjacent characters and.
Bayesian Filtering for Robot Localization
Exploiting Segmentation for Robust 3D Object Matching Michael Krainin, Kurt Konolige, and Dieter Fox.
Markov Localization & Bayes Filtering
Human Gesture Recognition Using Kinect Camera Presented by Carolina Vettorazzo and Diego Santo Orasa Patsadu, Chakarida Nukoolkit and Bunthit Watanapa.
From Bayesian Filtering to Particle Filters Dieter Fox University of Washington Joint work with W. Burgard, F. Dellaert, C. Kwok, S. Thrun.
A General Framework for Tracking Multiple People from a Moving Camera
Mapping and Localization with RFID Technology Matthai Philipose, Kenneth P Fishkin, Dieter Fox, Dirk Hahnel, Wolfram Burgard Presenter: Aniket Shah.
Visibility Graph. Voronoi Diagram Control is easy: stay equidistant away from closest obstacles.
#MOTION ESTIMATION AND OCCLUSION DETECTION #BLURRED VIDEO WITH LAYERS
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
Kevin Cherry Robert Firth Manohar Karki. Accurate detection of moving objects within scenes with dynamic background, in scenarios where the camera is.
A Progressive Tri-level Segmentation Approach for Topology-Change-Aware Video Matting Jinlong Ju 1, Jue Wang 3, Yebin Liu 1, Haoqian Wang 2, Qionghai Dai.
1 Markov Random Fields with Efficient Approximations Yuri Boykov, Olga Veksler, Ramin Zabih Computer Science Department CORNELL UNIVERSITY.
The 18th Meeting on Image Recognition and Understanding 2015/7/29 Depth Image Enhancement Using Local Tangent Plane Approximations Kiyoshi MatsuoYoshimitsu.
Peter Henry1, Michael Krainin1, Evan Herbst1,
Removing Moving Objects from Point Cloud Scenes Krystof Litomisky and Bir Bhanu International Workshop on Depth Image Analysis November 11, 2012.
Learning to Navigate Through Crowded Environments Peter Henry 1, Christian Vollmer 2, Brian Ferris 1, Dieter Fox 1 Tuesday, May 4, University of.
Robotics Club: 5:30 this evening
FREE-VIEW WATERMARKING FOR FREE VIEW TELEVISION Alper Koz, Cevahir Çığla and A.Aydın Alatan.
CSE-473 Project 2 Monte Carlo Localization. Localization as state estimation.
Motion Estimation using Markov Random Fields Hrvoje Bogunović Image Processing Group Faculty of Electrical Engineering and Computing University of Zagreb.
Journal of Visual Communication and Image Representation
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
Monte Carlo Localization for Mobile Robots Frank Dellaert 1, Dieter Fox 2, Wolfram Burgard 3, Sebastian Thrun 4 1 Georgia Institute of Technology 2 University.
SA-1 University of Washington Department of Computer Science & Engineering Robotics and State Estimation Lab Dieter Fox Stephen Friedman, Lin Liao, Benson.
Paper: A. Kapoor, H. Ahn, and R. Picard, “Mixture of Gaussian Processes for Combining Multiple Modalities,” MIT Media Lab Technical Report, Paper.
Learning and Removing Cast Shadows through a Multidistribution Approach Nicolas Martel-Brisson, Andre Zaccarin IEEE TRANSACTIONS ON PATTERN ANALYSIS AND.
Energy minimization Another global approach to improve quality of correspondences Assumption: disparities vary (mostly) smoothly Minimize energy function:
Learning color and locality cues for moving object detection and segmentation Yuan-Hao Lai Feng Liu and Michael Gleicher University of Wisconsin-Madison.
Learning Image Statistics for Bayesian Tracking Hedvig Sidenbladh KTH, Sweden Michael Black Brown University, RI, USA
Tracking Objects with Dynamics
Markov Random Fields with Efficient Approximations
SemanticFusion: Dense 3D Semantic Mapping with Convolutional Neural Networks Paper by John McCormac, Ankur Handa, Andrew Davison, and Stefan Leutenegger.
A New Approach to Track Multiple Vehicles With the Combination of Robust Detection and Two Classifiers Weidong Min , Mengdan Fan, Xiaoguang Guo, and Qing.
Parallel Integration of Video Modules
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
A Short Introduction to the Bayes Filter and Related Models
Analysis of Contour Motions
Principle of Bayesian Robot Localization.
Introduction to Object Tracking
Unsupervised Perceptual Rewards For Imitation Learning
Learning complex visual concepts
Presentation transcript:

Toward Object Discovery and Modeling via 3-D Scene Comparison Evan Herbst, Peter Henry, Xiaofeng Ren, Dieter Fox University of Washington; Intel Research Seattle 1

Overview Goal: learn about an environment by tracking changes in it over time Detect objects that occur in different places at different times 2 Handle textureless objects Avoid appearance/shape priors Represent a map with static + dynamic parts

Algorithm Outline Input: two RGB-D videos Mapping & reconstruction of each video Interscene alignment Change detection Spatial regularization Outputs: reconstructed static background; segmented movable objects 3

Scene Reconstruction Mapping based on RGB-D Mapping [Henry et al. ISER’10] Visual odometry, loop-closure detection, pose-graph optimization, bundle adjustment 4

Scene Reconstruction Mapping based on RGB-D Mapping [Henry et al. ISER’10] Surface representation: surfels 5

Scene Differencing Given two scenes, find parts that differ Surfaces in two scenes similar iff object doesn’t move Comparison at each surface point 6

Scene Differencing Given two scenes, find parts that differ Comparison at each surface point Start by globally aligning scenes 7 (2-D)(3-D)

Naïve Scene Differencing Easy algorithm: closest point within δ → same Ignores color, surface orientation Ignores occlusions 8

Model probability that a surface point moved Sensor readings z Expected measurement z* m ϵ {0, 1} Scene Differencing 9 z*z* z0z0 z1z1 z2z2 z3z3 frame 0 frame 10 frame 25 frame 49

Sensor Models 10 Model probability that a surface point moved Sensor readings z; expected measurement z* By Bayes, Two sensor measurement models With no expected surface: With expected surface:

Sensor Models Two sensor measurement models With expected surface Depth: uniform + exponential + Gaussian 1 Color: uniform + Gaussian Orientation: uniform + Gaussian 11 1 Thrun et al., Probabilistic Robotics, 2005 zd*zd*

Sensor Models Two sensor measurement models With expected surface Depth: uniform + exponential + Gaussian 1 Color: uniform + Gaussian Orientation: uniform + Gaussian With no expected surface Depth: uniform + exponential Color: uniform Orientation: uniform 12 1 Thrun et al., Probabilistic Robotics, 2005 zd*zd*

Example Result 13 Scene 1 Scene 2

Spatial Regularization 14 Points treated independently so far MRF to label each surfel moved or not moved Data term given by pointwise evidence Smoothness term: Potts, weighted by curvature

Spatial Regularization 15 Points treated independently so far MRF to label each surfel moved or not moved Scene 1 Scene 2 pointwise regularized

Experiments Trained MRF on four scenes (1.4M surfels) Tested on twelve scene pairs (8.0M surfels) 70% error reduction wrt max-class baseline 16 Count% % Total surfels8.0M1008.0M100 Moved surfels250k3 3 Errors250k355.5k0.7 False pos004.5k0.06 False neg250k351.0k0.64 BaselineOurs

Experiments Results: complex scene 17

Experiments Results: large object 18

Next steps All scenes in one optimization Model completion from many scenes Train more supervised object segmentation Conclusion Segment movable objects in 3-D using scene changes over time Represent a map as static + dynamic parts Extensible sensor model for RGB-D sensors 19

Using More Than 2 Scenes Given our framework, pretty easy to combine evidence from multiple scenes: w scene could be chosen to weight all scenes (rather than frames) equally, or upweight those taken under good lighting Other ways to subsample frames: as in keyframe selection in mapping 20

– Color, normal: uniform + Gaussian; mixing controlled by probability that beam hit expected surface First Sensor Model: Surface Didn’t Move Modeling sensor measurements: Depth: uniform + exponential + Gaussian * 21 * Fox et al., “Markov Localization…”, JAIR ‘99 zd*zd*

Experiments Trained MRF on four scenes (2.7 Msurfels) Tested on twelve scene pairs (8.0 Msurfels) 250k moved surfels; we get 4.5k FP, 51k FN 65% error reduction wrt max-class baseline Extract foreground segments as “objects” 22

Overview Many visits to same area over time Find objects by motion 23

(extra) Related Work Prob. Sensor models Depth only Depth & color, extra indep. Assumptions Static + dynamic maps In 2-d Usually not modeling objs 24

Spatial Regularization Pointwise only so far MRF to label each surfel moved or not moved Data term given by pointwise evidence Smoothness term: Potts, weighted by curvature 25

Depth-Dependent Color/Normal Model Modeling sensor measurements: Combine depth/color/normal: 26

Scene Reconstruction Mapping based on RGB-D Mapping [Henry et al. ISER’10] Surface representation: surfels 27