DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January 2000 1 Adaptive Intelligent Mobile Robotics Leslie Pack Kaelbling Artificial Intelligence.

Slides:



Advertisements
Similar presentations
Viktor Zhumatiya, Faustino Gomeza,
Advertisements

Reactive and Potential Field Planners
ROBOT RENDEZVOUS: 3 OR MORE ROBOTS USING 1-DIMENSIONAL SEARCH !!!!!!! !
Joshua Fabian Tyler Young James C. Peyton Jones Garrett M. Clayton Integrating the Microsoft Kinect With Simulink: Real-Time Object Tracking Example (
Artificial Intelligence Lecture 11. Computer Science Robotics & AI.
DARPA Mobile Autonomous Robot SoftwareMay Adaptive Intelligent Mobile Robotics William D. Smart, Presenter Leslie Pack Kaelbling, PI Artificial.
LCSLCS 18 September 2002DARPA MARS PI Meeting Intelligent Adaptive Mobile Robots Georgios Theocharous MIT AI Laboratory with Terran Lane and Leslie Pack.
Visual Navigation in Modified Environments From Biology to SLAM Sotirios Ch. Diamantas and Richard Crowder.
SA-1 Probabilistic Robotics Probabilistic Sensor Models Beam-based Scan-based Landmarks.
Bayesian Reinforcement Learning with Gaussian Processes Huanren Zhang Electrical and Computer Engineering Purdue University.
Practical Reinforcement Learning in Continuous Space William D. Smart Brown University Leslie Pack Kaelbling MIT Presented by: David LeRoux.
Experiences with an Architecture for Intelligent Reactive Agents By R. Peter Bonasso, R. James Firby, Erann Gat, David Kortenkamp, David P Miller, Marc.
Effective Reinforcement Learning for Mobile Robots William D. Smart & Leslie Pack Kaelbling* Mark J. Buller (mbuller) 14 March 2007 *Proceedings of IEEE.
Crowds Andrew Kaufman Michael Welsman-Dinelle. What is a crowd? A group of agents performing actions. Agents can affect each other. Agent actions may.
Octopus Arm Mid-Term Presentation Dmitry Volkinshtein & Peter Szabo Supervised by: Yaki Engel.
Reinforcement Learning: Learning to get what you want... Sutton & Barto, Reinforcement Learning: An Introduction, MIT Press 1998.
High Speed Obstacle Avoidance using Monocular Vision and Reinforcement Learning Jeff Michels Ashutosh Saxena Andrew Y. Ng Stanford University ICML 2005.
Simultaneous Localization and Map Building System for Prototype Mars Rover CECS 398 Capstone Design I October 24, 2001.
Behavior- Based Approaches Behavior- Based Approaches.
A Robust Layered Control System for a Mobile Robot Rodney A. Brooks Presenter: Michael Vidal.
Learning Programs Danielle and Joseph Bennett (and Lorelei) 4 December 2007.
CS 188: Artificial Intelligence Fall 2009 Lecture 19: Hidden Markov Models 11/3/2009 Dan Klein – UC Berkeley.
Exploration in Reinforcement Learning Jeremy Wyatt Intelligent Robotics Lab School of Computer Science University of Birmingham, UK
Robert M. Saltzman © DS 851: 4 Main Components 1.Applications The more you see, the better 2.Probability & Statistics Computer does most of the work.
8/9/20151 DARPA-MARS Kickoff Adaptive Intelligent Mobile Robots Leslie Pack Kaelbling Artificial Intelligence Laboratory MIT.
Reinforcement Learning in the Presence of Hidden States Andrew Howard Andrew Arnold {ah679
Autonomous Unmanned Ground Vehicle Navigation: Present and Future Larry Jackel DARPA IPTO / TTO darpatech2004/
Behavior Based Robotics: A Wall Following Behavior Arun Mahendra - Dept. of Math, Physics & Engineering, Tarleton State University Mentor: Dr. Mircea Agapie.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; March Adaptive Intelligent Mobile Robotics Leslie Pack Kaelbling Artificial Intelligence.
Zhiyong Wang In cooperation with Sisi Zlatanova
REINFORCEMENT LEARNING LEARNING TO PERFORM BEST ACTIONS BY REWARDS Tayfun Gürel.
ADVANCED CLASSIFICATION TECHNIQUES David Kauchak CS 159 – Fall 2014.
Leslie Luyt Supervisor: Dr. Karen Bradshaw 2 November 2009.
CS206Evolutionary Robotics Anatomy of an evolutionary robotics experiment: 1.Create a task environment. 2.Create the robot. 3.Create the robot’s brain,
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
Study on Genetic Network Programming (GNP) with Learning and Evolution Hirasawa laboratory, Artificial Intelligence section Information architecture field.
Natural Actor-Critic Authors: Jan Peters and Stefan Schaal Neurocomputing, 2008 Cognitive robotics 2008/2009 Wouter Klijn.
1 IE 607 Heuristic Optimization Particle Swarm Optimization.
Spatio-Temporal Case-Based Reasoning for Behavioral Selection Maxim Likhachev and Ronald Arkin Mobile Robot Laboratory Georgia Tech.
September Bound Computation for Adaptive Systems V&V Giampiero Campa September 2008 West Virginia University.
Reinforcement Learning Control with Robust Stability Chuck Anderson, Matt Kretchmar, Department of Computer Science, Peter Young, Department of Electrical.
DARPA ITO/MARS Project Update Vanderbilt University A Software Architecture and Tools for Autonomous Robots that Learn on Mission K. Kawamura, M. Wilkes,
Curiosity-Driven Exploration with Planning Trajectories Tyler Streeter PhD Student, Human Computer Interaction Iowa State University
Evolving the goal priorities of autonomous agents Adam Campbell* Advisor: Dr. Annie S. Wu* Collaborator: Dr. Randall Shumaker** School of Electrical Engineering.
Project by: Cirill Aizenberg, Dima Altshuler Supervisor: Erez Berkovich.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
A Roadmap towards Machine Intelligence
Selection of Behavioral Parameters: Integration of Case-Based Reasoning with Learning Momentum Brian Lee, Maxim Likhachev, and Ronald C. Arkin Mobile Robot.
Learning Momentum: Integration and Experimentation Brian Lee and Ronald C. Arkin Mobile Robot Laboratory Georgia Tech Atlanta, GA.
Evolving Cutting Horse and Sheepdog Behavior on a Simulated Flock Chris Beacham.
Artificial Intelligence: Research and Collaborative Possibilities a presentation by: Dr. Ernest L. McDuffie, Assistant Professor Department of Computer.
Particle Swarm Optimization (PSO)
Chapter 14. Active Vision for Goal-Oriented Humanoid Robot Walking (2/2) in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from.
A field of study that encompasses computational techniques for performing tasks that require intelligence when performed by humans. Simulation of human.
MIT Artificial Intelligence Laboratory — Research Directions Intelligent Agents that Learn Leslie Pack Kaelbling.
Robot Intelligence Technology Lab. 10. Complex Hardware Morphologies: Walking Machines Presented by In-Won Park
Evolving robot brains using vision Lisa Meeden Computer Science Department Swarthmore College.
NTT-MIT Collaboration Meeting, 2001Leslie Pack Kaelbling 1 Learning in Worlds with Objects Leslie Pack Kaelbling MIT Artificial Intelligence Laboratory.
Crowds (and research in computer animation and games)
Learning Fast and Slow John E. Laird
Chapter 11: Artificial Intelligence
Done Done Course Overview What is AI? What are the Major Challenges?
Reinforcement learning (Chapter 21)
Deep reinforcement learning
Crowds (and research in computer animation and games)
Networks of Autonomous Unmanned Vehicles
Continous-Action Q-Learning
Area Coverage Problem Optimization by (local) Search
Presentation transcript:

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Adaptive Intelligent Mobile Robotics Leslie Pack Kaelbling Artificial Intelligence Laboratory MIT

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Progress to Date Erik the Red Video game environment Optical flow implementation Fast bootstrapped reinforcement learning

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Erik the Red RWI B21 robot camera, sonars, laser range-finder, infrareds 3 Linux machines ported our framework for writing debuggable code

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Erik the Red

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Crystal Space Public-domain video-game environment complex graphics other agents highly modifiable

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Crystal Space

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Optical Flow Get range information visually by computing optical flow field nearer objects cause flow of higher magnitude expansion pattern means you’re going to hit rate of expansion tells you when elegant control laws based on center and rate of expansion (derived from human and fly behavior)

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Optical Flow in Crystal Space

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Making RL Really Work Typical RL methods require far too much data to be practical in an online setting. Address the problem by strong generalization techniques using human input to bootstrap

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January JAQL Learning a value function in a continuous state and action space based on locally weighted regression (fancy version of nearest neighbor) algorithm knows what it knows use meta-knowledge to be conservative about dynamic-programming updates

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Incorporating Human Input Humans can help a lot, even if they can’t perform the task very well. Provide some initial successful trajectories through the space Trajectories are not used for supervised learning, but to guide the reinforcement-learning methods through useful parts of the space Learn models of the dynamics of the world and of the reward structure Once learned models are good, use them to update the value function and policy as well.

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Simple Experiment The “hill-car” problem in two continuous dimensions Regular RL methods take thousands of trials to learn a reasonable policy JAQL takes 11 inefficient but eventually successful trails generated by humans to get 80% performance 10 more subsequent trials generate high quality performance in the whole space

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Success Percentage

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Trial Length (200 max) 54-step optimum

DARPA Mobile Autonomous Robot SoftwareLeslie Pack Kaelbling; January Next Steps Implement optical-flow control algorithms on robot Apply RL techniques to tune parameters in control algorithms on robot in real time corridor following using sonar and laser obstacle avoidance using optical flow Build highly complex simulated environment Integrate planning and learning in multi-layer system