Robot Vision: Multi-sensor Reconnaissance. Overview An individual robot can develop an interpretation about its environment. Groups of robots can combine.

Slides:



Advertisements
Similar presentations
1 ECE 495 – Integrated System Design I Introduction to Image Processing ECE 495, Spring 2013.
Advertisements

Bayesian Reconstruction of 3D Human Motion from Single-Camera Video
3D Head Mesh Data Stereo Vision Active Stereo 3D Reconstruction 3dMD System 1.
Exploration of bump, parallax, relief and displacement mapping
Multiview Reconstruction. Why More Than 2 Views? BaselineBaseline – Too short – low accuracy – Too long – matching becomes hard.
CS 376b Introduction to Computer Vision 04 / 21 / 2008 Instructor: Michael Eckmann.
VisHap: Guangqi Ye, Jason J. Corso, Gregory D. Hager, Allison M. Okamura Presented By: Adelle C. Knight Augmented Reality Combining Haptics and Vision.
Caught in Motion By: Eric Hunt-Schroeder EE275 – Final Project - Spring 2012.
Nathan Johnson1 Background Subtraction Various Methods for Different Inputs.
Vision Based Control Motion Matt Baker Kevin VanDyke.
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
Tracking Multiple Occluding People by Localizing on Multiple Scene Planes Professor :王聖智 教授 Student :周節.
Object Inter-Camera Tracking with non- overlapping views: A new dynamic approach Trevor Montcalm Bubaker Boufama.
Robust Object Tracking via Sparsity-based Collaborative Model
Multiple People Detection and Tracking with Occlusion Presenter: Feifei Huo Supervisor: Dr. Emile A. Hendriks Dr. A. H. J. Stijn Oomes Information and.
Shape from Contours and Multiple Stereo A Hierarchical, Mesh-Based Approach Hendrik Kück, Wolfgang Heidrich, Christian Vogelgsang.
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
A Study of Approaches for Object Recognition
Shadow Silhouette Maps Pradeep Sen, Mike Cammarano, Pat Hanrahan Stanford University.
1 Image-Based Visual Hulls Paper by Wojciech Matusik, Chris Buehler, Ramesh Raskar, Steven J. Gortler and Leonard McMillan [
Uncalibrated Geometry & Stratification Sastry and Yang
Multi-view stereo Many slides adapted from S. Seitz.
Fast Illumination-invariant Background Subtraction using Two Views: Error Analysis, Sensor Placement and Applications Ser-Nam Lim, Anurag Mittal, Larry.
The plan for today Camera matrix
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
Tracking Video Objects in Cluttered Background
Augmented Reality: Object Tracking and Active Appearance Model
Background Subtraction for Urban Traffic Monitoring using Webcams Master Graduation Project Final Presentation Supervisor: Rein van den Boomgaard Mark.
Multi-camera Video Surveillance: Detection, Occlusion Handling, Tracking and Event Recognition Oytun Akman.
Presented by Pat Chan Pik Wah 28/04/2005 Qualifying Examination
Advanced Computer Vision Structure from Motion. Geometric structure-from-motion problem: using image matches to estimate: The 3D positions of the corresponding.
Video Trails: Representing and Visualizing Structure in Video Sequences Vikrant Kobla David Doermann Christos Faloutsos.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
1 Video Surveillance systems for Traffic Monitoring Simeon Indupalli.
College of Engineering and Science Clemson University
Research & Innovation 1 An Industry Perspective on VVG Research Oliver Grau BBC Research & Innovation VVG SUMMER SCHOOL '07.
Image-Based Visual Hulls Wojciech Matusik Chris Buehler Ramesh Raskar Steven Gortler Leonard McMillan Presentation by: Kenton McHenry.
Optimal Placement and Selection of Camera Network Nodes for Target Localization A. O. Ercan, D. B. Yang, A. El Gamal and L. J. Guibas Stanford University.
Shadows Computer Graphics. Shadows Shadows Extended light sources produce penumbras In real-time, we only use point light sources –Extended light sources.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
COMP 175: Computer Graphics March 24, 2015
Exploitation of 3D Video Technologies Takashi Matsuyama Graduate School of Informatics, Kyoto University 12 th International Conference on Informatics.
Course 9 Texture. Definition: Texture is repeating patterns of local variations in image intensity, which is too fine to be distinguished. Texture evokes.
Vision-based human motion analysis: An overview Computer Vision and Image Understanding(2007)
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Vehicle Segmentation and Tracking From a Low-Angle Off-Axis Camera Neeraj K. Kanhere Committee members Dr. Stanley Birchfield Dr. Robert Schalkoff Dr.
Figure ground segregation in video via averaging and color distribution Introduction to Computational and Biological Vision 2013 Dror Zenati.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
Tutorial Visual Perception Towards Computer Vision
3d Pose Detection Used by Kinect
Hank Childs, University of Oregon Volume Rendering, pt 1.
Ray Tracing Fall, Introduction Simple idea  Forward Mapping  Natural phenomenon infinite number of rays from light source to object to viewer.
By Naveen kumar Badam. Contents INTRODUCTION ARCHITECTURE OF THE PROPOSED MODEL MODULES INVOLVED IN THE MODEL FUTURE WORKS CONCLUSION.
875 Dynamic Scene Reconstruction
ECE 638: Principles of Digital Color Imaging Systems Lecture 4: Chromaticity Diagram.
1Ellen L. Walker 3D Vision Why? The world is 3D Not all useful information is readily available in 2D Why so hard? “Inverse problem”: one image = many.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Occlusion Tracking Using Logical Models Summary. A Variational Partial Differential Equations based model is used for tracking objects under occlusions.
Computer Vision Computer Vision based Hole Filling Chad Hantak COMP December 9, 2003.
Zhaoxia Fu, Yan Han Measurement Volume 45, Issue 4, May 2012, Pages 650–655 Reporter: Jing-Siang, Chen.
Presented by 翁丞世  View Interpolation  Layered Depth Images  Light Fields and Lumigraphs  Environment Mattes  Video-Based.
Motion Estimation of Moving Foreground Objects Pierre Ponce ee392j Winter March 10, 2004.
Robot Vision SS 2011 Matthias Rüther / Matthias Straka 1 ROBOT VISION Lesson 7: Volumetric Object Reconstruction Matthias Straka.
Vehicle Segmentation and Tracking in the Presence of Occlusions
Vehicle Segmentation and Tracking from a Low-Angle Off-Axis Camera
Advanced Computer Vision
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Occlusion and smoothness probabilities in 3D cluttered scenes
Report 2 Brandon Silva.
Presentation transcript:

Robot Vision: Multi-sensor Reconnaissance

Overview An individual robot can develop an interpretation about its environment. Groups of robots can combine their resources to develop an improved, collective interpretation. Such a “swarm” is reliant on centralized processing to refine the collective interpretation. The swarm has more insight than solitary individuals.

Single-sensor Vision Our setup involves numerous robots, each with a single video camera. What can a single-camera robot see? A background image may be used to establish a baseline of irrelevant imagery. Anything that differs from the background is considered relevant – background subtraction. minusequals

Making Sense of Pixel Data Background subtraction and classical image processing algorithms can mark every pixel in every frame as foreground (interesting) or background (uninteresting). Conventional techniques can clean noise. Regions of contiguous foreground pixels are likely to constitute a single object: a “blob.” This assumption is false when one object occludes another. Later we will see a way around this.

Developing the Individual’s Interpretation Pixel-space blobs have measurable properties: color, shape, and location. Geometric properties (location, size) are relative to the camera’s perspective. A single robot can provide a list of objects with a bearing to each, but has no depth perception. Each robot’s interpretation may be combined to form a much stronger interpretation for the entire swarm.

Developing the Swarm’s Interpretation Each robot has a list of bearings to potential objects. This information can be visualized as rays originating from the robot’s location. The intersections of these rays represent potential objects in 3D space. Many intersections are bogus. Many intersections conflict with others – each ray can only correspond to one object.

Culling False Objects Algorithm: group compatible locations together. This yields disjoint sets of intersections that can coexist. The set with the most supporting evidence wins.

Combining the Individual and Swarm Interpretations At this point we have a set of objects with 3D locations. Individual robots can provide silhouettes of the objects. This information may be combined to create a 3D shape. Incorporating past history can strengthen our conclusions.

3D Hulls Each camera contributes a silhouette of an object, and a ray on which the silhouette lies. Projecting the silhouette along the ray forms a “cone”. The intersection of these cones carves a 3D solid; imagine pushing cookie cutters through space. The solid is guaranteed to enclose the true shape, but will be convex, i.e., ignores indentations. Such an upper bound is termed a hull. Hulls are typically represented as a mesh of triangles.

Applications of 3D Hulls The 2D silhouettes can be “painted” on the mesh. The solid can be rendered from any angle. 3D shape may be used to classify objects – threat assessment, for example. Meshes may be recorded for future use…

Establishing Object Tracks Object records from successive frames can be combined to establish a log of known objects. These “object tracks” can aid future processing, establishing a positive feedback loop: –Distinguishing between one large object, and two objects close to one another. –Past motion can predict where objects will be located, minimizing the occlusion problem mentioned earlier. –Past hulls can predict how an object’s silhouette will appear in each camera.

Summary Swarms of camera-equipped robots can collaborate to track and model objects in space. The swarm’s results are more concrete than any individual’s observations. Observation is passive and uses relatively few resources (weight, energy, money).