Pierre Sermanet¹·² Raia Hadsell¹ Jan Ben² Ayse Naz Erkan¹ Beat Flepp² Urs Muller² Yann LeCun¹ (1) Courant Institute of Mathematical Sciences, New York.

Slides:



Advertisements
Similar presentations
Project Title Here IEEE UCSD Overview Robo-Magellan is a robotics competition emphasizing autonomous navigation and obstacle avoidance over varied, outdoor.
Advertisements

IT253: Computer Organization
AI Pathfinding Representing the Search Space
ARCHITECTURES FOR ARTIFICIAL INTELLIGENCE SYSTEMS
Real-time, low-resource corridor reconstruction using a single consumer grade RGB camera is a powerful tool for allowing a fast, inexpensive solution to.
A vision-based system for grasping novel objects in cluttered environments Ashutosh Saxena, Lawson Wong, Morgan Quigley, Andrew Y. Ng 2007 Learning to.
Electrical & Computer Engineering Dept. University of Patras, Patras, Greece Evangelos Skodras Nikolaos Fakotakis.
1 E190Q – Project Introduction Autonomous Robot Navigation Team Member 1 Name Team Member 2 Name.
A Cloud-Assisted Design for Autonomous Driving Swarun Kumar Shyamnath Gollakota and Dina Katabi.
Abstract This project focuses on realizing a series of operational improvements for WPI’s unmanned ground vehicle Prometheus with the end goal of a winning.
Intelligent Ground Vehicle Competition 2006 Brigham Young University.
Autonomy for Ground Vehicles Status Report, July 2006 Sanjiv Singh Associate Research Professor Field Robotics Center Carnegie Mellon University.
A Self-Supervised Terrain Roughness Estimator for Off-Road Autonomous Driving David Stavens and Sebastian Thrun Stanford Artificial Intelligence Lab.
Tracking a moving object with real-time obstacle avoidance Chung-Hao Chen, Chang Cheng, David Page, Andreas Koschan and Mongi Abidi Imaging, Robotics and.
Brent Dingle Marco A. Morales Texas A&M University, Spring 2002
Chapter 13 Embedded Systems
X96 Autonomous Robot Design Review Saturday, March 13, 2004 By John Budinger Francisco Otibar Scott Ibara.
CS533 - Concepts of Operating Systems
DAMN : A Distributed Architecture for Mobile Navigation Julio K. Rosenblatt Presented By: Chris Miles.
X96 Autonomous Robot Design Review Thursday, May 6, 2004 By John Budinger Francisco Otibar.
TAXISAT PROJECT Low Cost GNSS and Computer Vision based data fusion solution for driverless vehicles Marc POLLINA
June 12, 2001 Jeong-Su Han An Autonomous Vehicle for People with Motor Disabilities by G. Bourhis, O.Horn, O.Habert and A. Pruski Paper Review.
Fuzzy control of a mobile robot Implementation using a MATLAB-based rapid prototyping system.
ROBOT LOCALISATION & MAPPING: MAPPING & LIDAR By James Mead.
Autonomous Unmanned Ground Vehicle Navigation: Present and Future Larry Jackel DARPA IPTO / TTO darpatech2004/
Zereik E., Biggio A., Merlo A. and Casalino G. EUCASS 2011 – 4-8 July, St. Petersburg, Russia.
Digital System Bus A bus in a digital system is a collection of (usually unbroken) signal lines that carry module-to-module communications. The signals.
Autonomous Surface Navigation Platform Michael Baxter Angel Berrocal Brandon Groff.
Chapter 5.4 Artificial Intelligence: Pathfinding.
Mobile Robot Control Architectures “A Robust Layered Control System for a Mobile Robot” -- Brooks 1986 “On Three-Layer Architectures” -- Gat 1998? Presented.
HRTC Meeting 12 September 2002, Vienna Smart Sensors Thomas Losert.
Behavior Based Robotics: A Wall Following Behavior Arun Mahendra - Dept. of Math, Physics & Engineering, Tarleton State University Mentor: Dr. Mircea Agapie.
Juhana Leiwo – Marco Torti.  Position and movement  Direction of acceleration (gravity) ‏  Proximity and collision sensing  3-dimensional spatial.
Ruslan Masinjila Aida Militaru.  Nature of the Problem  Our Solution: The Roaming Security Robot  Functionalities  General System View  System Design.
Robot Autonomous Perception Model For Internet-Based Intelligent Robotic System By Sriram Sunnam.
Autonomous Vehicles By: Rotha Aing. What makes a vehicle autonomous ? “Driverless” Different from remote controlled 3 D’s –Detection –Delivery –Data-Gathering.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
Group 99 Mobile Vision for Autonomous… Navigation and Reconnaissance Jay Silver Kevin Wortman Advised by Bill Ross.
Software Development Software Testing. Testing Definitions There are many tests going under various names. The following is a general list to get a feel.
Implementing Codesign in Xilinx Virtex II Pro Betim Çiço, Hergys Rexha Department of Informatics Engineering Faculty of Information Technologies Polytechnic.
Robotica Lecture 3. 2 Robot Control Robot control is the mean by which the sensing and action of a robot are coordinated The infinitely many possible.
Autonomous Guidance Navigation and Control Michael Gillham University of Kent SYSIASS Meeting University of Essex
Phong Le (EE) Josh Haley (CPE) Brandon Reeves (EE) Jerard Jose (EE)
1 Deterministic Collision-Free Communication Despite Continuous Motion ALGOSENSORS 2009 Saira Viqar Jennifer L. Welch Parasol Lab, Department of CS&E TEXAS.
1 Distributed and Optimal Motion Planning for Multiple Mobile Robots Yi Guo and Lynne Parker Center for Engineering Science Advanced Research Computer.
University of Windsor School of Computer Science Topics in Artificial Intelligence Fall 2008 Sept 11, 2008.
Behavior Control for Robotic Exploration of Planetary Surfaces Written by Erann Gat, Rajiv Desai, Robert Ivlev, John Loch and David P Miller Presented.
Institute for Computer Science VI Autonomous Intelligent Systems
Topological Path Planning JBNU, Division of Computer Science and Engineering Parallel Computing Lab Jonghwi Kim Introduction to AI Robots Chapter 9.
Learning to Navigate Through Crowded Environments Peter Henry 1, Christian Vollmer 2, Brian Ferris 1, Dieter Fox 1 Tuesday, May 4, University of.
Abstract A Structured Approach for Modular Design: A Plug and Play Middleware for Sensory Modules, Actuation Platforms, Task Descriptions and Implementations.
AN INTELLIGENT ASSISTANT FOR NAVIGATION OF VISUALLY IMPAIRED PEOPLE N.G. Bourbakis*# and D. Kavraki # #AIIS Inc., Vestal, NY, *WSU,
Spatiotemporal Saliency Map of a Video Sequence in FPGA hardware David Boland Acknowledgements: Professor Peter Cheung Mr Yang Liu.
Rajeev Rai Bhatia 2K7-MRCE-ME-040. Team: Rajeev R Bhatia Tushar Chugh Adeel Kidwai Mentor: Prof. M.S. Giri.
Visual Odometry for Ground Vehicle Applications David Nistér, Oleg Naroditsky, and James Bergen Sarnoff Corporation CN5300 Princeton, New Jersey
Artificial Intelligence in Game Design Lecture 8: Complex Steering Behaviors and Combining Behaviors.
1 The Role and Design of V2X Communications for Automated Driving Gaurav Bansal Toyota InfoTechnology Center, USA Mountain View, CA
Fast SLAM Simultaneous Localization And Mapping using Particle Filter A geometric approach (as opposed to discretization approach)‏ Subhrajit Bhattacharya.
Path Planning Based on Ant Colony Algorithm and Distributed Local Navigation for Multi-Robot Systems International Conference on Mechatronics and Automation.
Dynamic Mission Planning for Multiple Mobile Robots Barry Brumitt and Anthony Stentz 26 Oct, 1999 AMRS-99 Class Presentation Brian Chemel.
Heterogeneous Teams of Modular Robots for Mapping and Exploration by Grabowski et. al.
Yann LeCun Other Methods and Applications of Deep Learning Yann Le Cun The Courant Institute of Mathematical Sciences New York University
Computer Organization and Machine Language Programming CPTG 245
COGNITIVE APPROACH TO ROBOT SPATIAL MAPPING
Pursuit-Evasion Games with UGVs and UAVs
Autonomous Robots Key questions in mobile robotics What is around me?
Search-Based Footstep Planning
Robot Intelligence Kevin Warwick.
Managing Global System
Presentation transcript:

Pierre Sermanet¹·² Raia Hadsell¹ Jan Ben² Ayse Naz Erkan¹ Beat Flepp² Urs Muller² Yann LeCun¹ (1) Courant Institute of Mathematical Sciences, New York University (2) Net-Scale Technologies, Morganville, NJ 07751, USA SPEED-RANGE DILEMMAS FOR VISION-BASED NAVIGATION IN UNSTRUCTURED TERRAIN

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Outline Program and System overview Problem definition Architecture Results

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Overview: Program LAGR: Learning Applied to Ground Robots Demonstrate learning algorithms in unstructured outdoor robotics Vision-based only (passive), no expensive equipment Reach a GPS goal the fastest without any prior knowledge of location DARPA funded, 10 teams (Universities and companies), common platform Comparison to state-of-the-art CMU “baseline” software and other teams Monthly tests by DARPA in various unknown locations: Overview Problem Architecture Results Unstructured outdoor robotics is highly challenging due to wide diversity of environments (colors, shapes, sizes of obstacles, lighting and shadows, etc)  Conventional algorithms are unsuited, need for adaptability and learning SwRI, TX Ft. Belvoir, VA Ft. Belvoir, VA Hanover, NH

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Overview: Platform Constructor: CMU/NREC Vision based only: 2 stereo pairs of cameras (+ GPS for global navigation) 4 Linux machines linked by Gigabit ethernet: Two “eye” machines (dual core 2Ghz): Image processing “planner” machine (single core 2Ghz): Planning and control loop “controller” machine: Low level communication Maximum speed: 1.3m/s Proprietary CMU/NREC API to sensors and actuators Proprietary CMU/NREC “Baseline”: end-to-end navigation software (D*, etc) (not re-used) Overview Problem Architecture Results GPS Dual stereo cameras Bumper

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Overview: Philosophy Main goal: Demonstrate machine learning algorithms for long-range vision (RSS07). Supporting goal: Build a solid software platform for long-range vision and navigation: Robust and reliable Resistant to sensors imprecisions and failures Overview Problem Architecture Results Input image Stereo labels (short-range) Input: context-rich image windows Output: long-range labels Self-supervised learning using convolutional network:

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Overview: System Processing chain: Overview Problem Architecture Results Sensors (cameras) Input image Image processing Traversibility map Network transmission Actuators (wheels) Path planning Path Eye machine Planner machine Control Pose (GPS + IMU + wheels) Latency Frequency Note: Latency is not only tied to frequency but also to sensors latency, network, planning and actuators latency.

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Problem Important performance drop in local obstacle avoidance with too high latency and frequency: Overview Problem Architecture Results Artificially increasing latency and period almost linearly increases the number of crashes in obstacles Human expert drivers of the UPI Crusher vehicle reported a feedback latency of 400ms was the maximum for good remote driving. How to guarantee good performance with increasing complexity introduced by sophisticated long-range vision modules? When does processing speed prevails over vision range, and vice- versa? Performance Test of July 2006, Holmdel Park, NJ

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Problem: Delays Overview Problem Architecture Results fixed Latency and frequency determine performance, but latency is actually composed of 3 types of latencies or “delays”: 1.Sensors/Actuators latency + LAGR API latency: Images are already 190ms old when made available to image processing 2.Processing latency 3.Robot’s dynamics latency (inertia + acceleration/deceleration): 1.5sec (worst case) between a wheel command and actual desired speed (1) and (3) are relatively high on the LAGR platform and must be caught up to and taken in account by (2).

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Problem: Solutions to delays Overview Problem Architecture Results To account for sensors and processing latencies (1) and (2): a. Reduce processing time. b. Estimate delays between path planning and actuation. c. Place traversibility maps according to delays before and after path planning. To account for dynamics latencies (3): d. Modeling or record robot’s dynamics.  All (a), (b), (c) and (d) solutions are part of the global solution presented in the results section, but here we will only describe a successful architecture for (a)

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Architecture Overview Problem Architecture Results Idea: Wagner et al.¹ showed that a walking human gazes more frequently close by than far away:  need higher frequency closer than far away Close by obstacles move toward robot faster than far obstacles:  need lower latency closer than far away To satisfy those requirements, short and long range vision must be separated into 2 parallel and independent OD modules: “Fast-OD”: processing has to be fast, vision is not necessarily long-range. “Far-OD”: vision has to be long-range, processing can be slower. How to make Fast-OD fast?  Simple processing and reduced input resolution.  Can we reduce resolution without reducing performance? ¹ M. Wagner, J. C. Baird, andW. Barbaresi. The locus of environmental attention. J. of Environmental Psychology, 1: , 1980.

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Architecture: Fast-OD // Far-OD Sensors (cameras) High res input image (320x240 or 512x384) Advanced Image processing Traversibility map (5 to >30m) Network transmission Actuators (wheels) Path planning Path Eye machine Planner machine Control Latency: 700ms Frequency: 3Hz Latency: 250ms Frequency: 10Hz Pose (GPS + IMU + wheels) Low res input image (160x120) Simple Image processing Traversibility map (0 to 5m) Map merging Overview Problem Architecture Results Fast-OD Far-OD

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Architecture: Implementation notes Overview Problem Architecture Results CPU cycles: All cycles must be given to Fast-OD when it runs to guarantee low latency. Different solutions are: 1.Use real-time OS and give high priority to Fast-OD. 2.With regular OS, give Fast-OD control of Far-OD:  Fast-OD pauses Far-OD, runs, then sleeps for a bit and resume Far-OD. 3.Use dual-core CPU. Map merging: Fast and Far maps are merged together before planning according to their respective poses. 2-step planning: This architecture makes it easier to separate the different planning algorithms suited for short and long range: Fast-OD planning happens in Cartesian space and takes robot dynamics in account (more important in short range) Far-OD planning happens in image space and uses regular path planning. Dynamics planning: Cartesian space Long-range planning: Image space 10m 5m infinity 10m

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Results: Timing measures Overview Problem Architecture Results Fast-od actuation latency 250ms Fast-od sensors latency 190ms Fast-od period (frequency) 100ms (10Hz) Far-od period (frequency) 370ms (2-3Hz) Far-od actuation latency 700ms

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Global-map Vehicle-map Results Overview Problem Architecture Results Short and long range navigation test: 1.1 st obstacle appears quickly and suddenly to robot  testing short range navigation 2.Cul-de-sac  testing “long” range navigation  Parallel architecture is consistently better at short and long range navigation than series architecture or FAST-OD only. Note: Here Fast-od has 5m radius and Far-od 15m radius.

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Results: More recent results Overview Problem Architecture Results Fast-od + Far-od in parallel: Short-range navigation consistently successful: 0 collision over >5 runs Finish run in about 16sec along shortest path Fast-od: 10Hz & 250ms & 3meters range Far-od: 3Hz & 700ms & 30meters range Fast-od + Far-od in series: Short-range navigation consistently failing: > 2 collisions over >5 runs Finish run in >40sec along longer path Fast-od/Far-od: 3Hz & 700ms & 3m/30m range (frequency is acceptable but latency is too high) Video 1: collision-free bucket maze Video 2: collision-free bucket maze Videos 3,4,5: obstacle collisions due to high latency and period.

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Results: More recent results Overview Problem Architecture Results Video 6: Natural obstacles Video 7: Tight maze of artificial obstacles Fast-od + Far-od in parallel: Short-range navigation consistently successful: 0 collision over >5 runs Fast-od: 10Hz & 250ms & 3meters range Far-od: 3Hz & 700ms & 30meters range Note: long-range planning is off, i.e. Far-od is processing but ignored. Only short-range navigation was tested here.

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Results: Moving obstacles Overview Problem Architecture Results Video 8: Fast moving obstacle Detects and avoids moving obstacles consistently.

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Results: Beating humans Overview Problem Architecture Results Video 9: Experienced human driver Autonomous short-range navigation is consistently better than inexperienced human drivers and equal or better than experienced human drivers. (driving with only robot’s images would be even harder for a human)

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Results: Processing Speed - Vision Range dilemma Overview Problem Architecture Results We showed that processing speed prevails over vision range for short range navigation, whereas vision range prevails over speed for long range navigation. Only 3m vision range were necessary to build a collision-free short range navigation for a 1.3m/s non-holonomic vehicle: Vehicle’s worst-case stopping delay: 1.0 sec. System’s worst-case reaction time: 0.25 sec. latency sec period  Worst-case reaction and stopping delay: 1.35 sec., (or 1.75m)  Only 1.0 sec. anticipation necessary in addition to worst-case reaction and stopping delay. A vision range of 15m with high latency and lower frequency consistently improved the long range navigation in parallel to the short range module.

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Summary Overview Problem Architecture Results We showed that both latency and frequency are critical in vision-based systems because of higher processing times. A simple and very low resolution OD in parallel with a high resolution OD proved to increase greatly the performance of a short and long range vision- based autonomous navigation system over commonly used higher resolution and sequential approaches: Processing speed prevails over range in short-range navigation and only 1.0 sec. additional anticipation to dynamics and processing delays was necessary. Additional key concepts such as dynamics modeling must be implemented to build a complete end-to-end successful system. A robust collision-free navigation platform, dealing with moving obstacles and beating humans, was successfully built and is able to leave enough CPU cycles available for computationally expensive algorithms.

Intelligent Autonomous Vehicles IAV 2007, Toulouse, France Pierre Sermanet September 4 th, /21 Questions?