Apprenticeship Learning for Motion Planning with Application to Parking Lot Navigation Presented by: Paul Savickas.

Slides:



Advertisements
Similar presentations
Vision-based Motion Planning for an Autonomous Motorcycle on Ill-Structured Road Reporter :鄒嘉恆 Date : 08/31/09.
Advertisements

Miller Pipeline Corp. DRIVER SAFETY TRAINING. SECTION 1 Introduction.
Decision Making and control for automotive safety Mohammad Ali.
Optimizing Flocking Controllers using Gradient Descent
Adam Coates, Pieter Abbeel, and Andrew Y. Ng Stanford University ICML 2008 Learning for Control from Multiple Demonstrations TexPoint fonts used in EMF.
Apprenticeship Learning for Robotic Control, with Applications to Quadruped Locomotion and Autonomous Helicopter Flight Pieter Abbeel Stanford University.
International Conference on Automatic Face and Gesture Recognition, 2006 A Layered Deformable Model for Gait Analysis Haiping Lu, K.N. Plataniotis and.
STANFORD Hierarchical Apprenticeship Learning with Application to Quadruped Locomotion J. Zico Kolter, Pieter Abbeel, Andrew Y. Ng Goal Initial Position.
Jessie Byrne L. Keliher COMP 1631 January 27 th 2011.
Stanford hci group / stanford · 14 January 2008 Designing Interface Alternatives with Parallel Exploration and Runtime Tuning.
Autonomy for Ground Vehicles Status Report, July 2006 Sanjiv Singh Associate Research Professor Field Robotics Center Carnegie Mellon University.
A.G.I.L.E Team Members: Brad Ramsey Derek Rodriguez Dane Wielgopolan Project Managers: Dr. Joel Schipper Dr. James Irwin Autonomously Guided Intelligent.
A Self-Supervised Terrain Roughness Estimator for Off-Road Autonomous Driving David Stavens and Sebastian Thrun Stanford Artificial Intelligence Lab.
Pieter Abbeel and Andrew Y. Ng Apprenticeship Learning via Inverse Reinforcement Learning Pieter Abbeel Stanford University [Joint work with Andrew Ng.]
Link to Presentation Not Easy Old House.
Apprenticeship Learning by Inverse Reinforcement Learning Pieter Abbeel Andrew Y. Ng Stanford University.
Apprenticeship Learning by Inverse Reinforcement Learning Pieter Abbeel Andrew Y. Ng Stanford University.
Pieter Abbeel and Andrew Y. Ng Apprenticeship Learning via Inverse Reinforcement Learning Pieter Abbeel Stanford University [Joint work with Andrew Ng.]
Grip on speed Polis Conference 'Innovation in transport for sustainable cities & regions’ november 2010 A multi-stakeholder approach to road safety.
Geoencryption Demonstration Di Qiu, Sherman Lo, Per Enge August
Pieter Abbeel and Andrew Y. Ng Apprenticeship Learning via Inverse Reinforcement Learning Pieter Abbeel and Andrew Y. Ng Stanford University.
Pieter Abbeel and Andrew Y. Ng Reinforcement Learning and Apprenticeship Learning Pieter Abbeel and Andrew Y. Ng Stanford University.
Deon Blaauw Modular Robot Design University of Stellenbosch Department of Electric and Electronic Engineering.
SKILL LEARNING Stages of Learning: page Types of Practice: page
ROBOT LOCALISATION & MAPPING: NAVIGATION Ken Birbeck.
Daphne Koller Stanford University Learning to improve our lives.
DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; March Adaptive Intelligent Mobile Robotics Leslie Pack Kaelbling Artificial Intelligence.
Artificial Intelligence: Prospects for the 21 st Century Henry Kautz Department of Computer Science University of Rochester.
Robotics. Foundational Methods Advanced Applications.
Autonomous Vehicles By: Rotha Aing. What makes a vehicle autonomous ? “Driverless” Different from remote controlled 3 D’s –Detection –Delivery –Data-Gathering.
Information Visualization DH /02/21 Proposal for Project 1 Title Cool Photo of Student 1 Student 1 Cool Photo of Student 2 Student.
Towards Automatic Driving Xiaokai He – January 20 th, 2010.
Google Self-Driving Cars By: Shiv Patel ITMG /16/13.
Parameter/State Estimation and Trajectory Planning of the Skysails flying kite system Jesus Lago, Adrian Bürger, Florian Messerer, Michael Erhard Systems.
Velodyne Lidar Sensor. Plane Detection in a 3D environment using a Velodyne Lidar Jacoby Larson UCSD ECE 172.
Accident Scene Safety Module 1 – Vehicle Safety Section 1 - Driving Safety.
Introduction to Level Set Methods: Part II
Urban Challenge Sponsored by DARPA November 3 rd, 2007.
Learning to Navigate Through Crowded Environments Peter Henry 1, Christian Vollmer 2, Brian Ferris 1, Dieter Fox 1 Tuesday, May 4, University of.
Texting and Driving Statistics.
Autonomous Vehicle Instructor Dr. Dongchul Kim Presented By Harish Kumar Gudipati.
DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Adaptive Intelligent Mobile Robotics Leslie Pack Kaelbling Artificial Intelligence.
Pathfinding Algorithms for Mutating Weight Graphs Haitao Mao Computer Systems Lab
CIS 2011 rkala.99k.org 1 st September, 2011 Planning of Multiple Autonomous Vehicles using RRT Rahul Kala, Kevin Warwick Publication of paper: R. Kala,
Autonomous Vehicle Initiatives at Princeton University Alain L. Kornhauser Professor, Operations Research & Financial Engineering Director, Transportation.
CONDITIONALS. FIRST CONDITIONAL If the weather is nice, we will go for a walk. If you don’t apologize, she will never trust you again. The first conditional.
Was it worth it? Texting & Driving By: Marquavius Ward.
Optimal Acceleration and Braking Sequences for Vehicles in the Presence of Moving Obstacles Jeff Johnson, Kris Hauser School of Informatics and Computing.
Reinforcement learning (Chapter 21)
Learning optimal behavior
Case Study Autonomous Cars 9/21/2018.
PAVE & the DARPA Challenges
E190Q – Project Introduction Autonomous Robot Navigation
Autonomous Vehicle Simulations
Daniel Brown and Scott Niekum The University of Texas at Austin
Stereo Vision Applications
Project Overview Introduction Frame Build Motion Power Control Sensors
Using Engineering Design and Learning Science to Create New Lessons
Nike Training Club App By: Kacy Maska.
TUGS Jason Higuchi && Julia Yefimenko && Raudel mayorga
Apprenticeship Learning via Inverse Reinforcement Learning
Case Study Autonomous Cars 1/14/2019.
Passing other Vehicles
MOTOR VEHICLE DRIVING PRACTICES.
Building Your Robocar Day 1.
Using Engineering Design and Learning Science to Create New Lessons
Human Powered Vehicle (HPV)
Race Questions: Question 1
Li Li, Zhu Li, Vladyslav Zakharchenko, Jianle Chen, Houqiang Li
Presentation transcript:

Apprenticeship Learning for Motion Planning with Application to Parking Lot Navigation Presented by: Paul Savickas

Outline Motivation Background The Problem Methods and Model Results Discussion

Motivation In 2005: ~6,420,000 Auto Accidents in the US Total financial cost >$230 Billion Injuries ~2.9 Million 42,636 Deaths – 115 per day – 1 every 13 minutes Source:

Background DARPA Urban Challenge Stanford Racing Team Junior

Stanford’s “Junior” oltw&NR=1 oltw&NR=1

The Problem Motion/Path planning algorithms are complex – Many parameters – Hand tuning required How can we simplify this process?

Methods Path-Planning and Optimization – Sequence of states – Potential-field terms – Weights

Apprenticeship Learning Learn parameters from expert demonstration Useful when no reward function available Example: Teaching a person to drive

Model Parameters – Length – Length (Reverse) – Direction Changes – Proximity to Obstacles – Smoothness – Parking Lot Conventions – Lane Conventions

Parameter Tuning Switching DirectionsPenalize offroad Distance from lane Penalize Switching Follow Directions

Experiment Human Driver demonstration – “Nice” – “Sloppy” – “Reverse-allowed” Autonomous Navigation using learned parameters

Results – “Nice”

Results – “Sloppy”

Results – “Reverse-allowed”

Results – Interesting Anomalies

Discussion How do these results apply to other problems? Determination of parameters is still an issue. Why would we want to drive “sloppy”?.. Where do we go from here?

Video DARPA Urban Challenge bJhA&feature=related bJhA&feature=related Junior’s Results h8 h8