Artificial Curiosity Tyler Streeter www.vrac.iastate.edu/~streeter.

Slides:



Advertisements
Similar presentations
Approaches, Tools, and Applications Islam A. El-Shaarawy Shoubra Faculty of Eng.
Advertisements

Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Learning School of Computing, University of Leeds, UK AI23 – 2004/05 – demo 2.
IJCNN 2013 Panel in Higher Emotion 1 Emotion in Brain-Inspired Developmental Networks Juyang (John) Weng Computer Sci., Neurosci., Cognitive Sci. Michigan.
Finding aesthetic pleasure on the edge of chaos: A proposal for robotic creativity Ron Chrisley COGS Department of Informatics University of Sussex Workshop.
Reinforcement Learning and the Reward Engineering Principle Daniel Dewey AAAI Spring.
Developmental Artificial Intelligence 1 er April t oliviergeorgeon.com1/31.
1.Data categorization 2.Information 3.Knowledge 4.Wisdom 5.Social understanding Which of the following requires a firm to expend resources to organize.
ROBOT BEHAVIOUR CONTROL SUCCESSFUL TRIAL OF MARKERLESS MOTION CAPTURE TECHNOLOGY Student E.E. Shelomentsev Group 8Е00 Scientific supervisor Т.V. Alexandrova.
Chapter 10 Artificial Intelligence © 2007 Pearson Addison-Wesley. All rights reserved.
Intrinsically Motivated Hierarchical Reinforcement Learning Andrew Barto Autonomous Learning Lab Department of Computer Science University of Massachusetts.
Dopamine, Uncertainty and TD Learning CNS 2004 Yael Niv Michael Duff Peter Dayan Gatsby Computational Neuroscience Unit, UCL.
Introduction to Neural Network Justin Jansen December 9 th 2002.
Cooperative Q-Learning Lars Blackmore and Steve Block Expertness Based Cooperative Q-learning Ahmadabadi, M.N.; Asadpour, M IEEE Transactions on Systems,
Mental Development and Representation Building through Motivated Learning Janusz A. Starzyk, Ohio University, USA, Pawel Raif, Silesian University of Technology,
Intelligent Agents revisited.
IJCNN, International Joint Conference on Neural Networks, San Jose 2011 Pawel Raif Silesian University of Technology, Poland, Janusz A. Starzyk Ohio University,
A.M. Florea, Cognitive systems, COST Action IC0801 – WG1, 15 December, Ayia Napa, Cyprus.
Curious Reconfigurable Robots Kathryn Merrick ITEE Seminar – Friday 29 th August Collaborators: Dayne Schmidt, Elanor Huntington.
Fourth International Symposium on Neural Networks (ISNN) June 3-7, 2007, Nanjing, China Online Dynamic Value System for Machine Learning Haibo He, Stevens.
Brain, Mind, Body and Society: Controllability and Uncontrollability in Robotics Motomu SHIMODA, PhD. Kyoto Women’s University.
Curious Characters in Multiuser Games: A Study in Motivated Reinforcement Learning for Creative Behavior Policies * Mary Lou Maher University of Sydney.
Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Cognitive Robots © 2014, SNU CSE Biointelligence Lab.,
A Software Architecture and Tools for Autonomous Robots that Learn on Mission K. Kawamura, M. Wilkes, A. Peters, D. Gaines* Vanderbilt University Center.
Developmental Mechanisms for Life-Long Autonomous Learning in Robots Pierre-Yves Oudeyer Project-Team INRIA-ENSTA-ParisTech FLOWERS
Artificial Intelligence By Ryan Shoultes & Jeremy Creighton.
Motivational Processes Affecting Learning
Com3240 Adaptive Intelligence Epigenetic Robotics Amanda Sharkey Reinforcement Learning Eleni Vasilaki.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
Learning analytics is the use of intelligent data, learner-produced data, and analysis models to discover information and social connections, and to predict.
Is the technology developmentally appropriate? Evaluation Tool Educational Value Is the content based on research/standards? Does the software follow the.
Integrating high- and low-level Expectations in Deliberative Agents Michele Piunti - Institute of.
Beyond Gazing, Pointing, and Reaching A Survey of Developmental Robotics Authors: Max Lungarella, Giorgio Metta.
Synthetic Cognitive Agent Situational Awareness Components Sanford T. Freedman and Julie A. Adams Department of Electrical Engineering and Computer Science.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 11: Artificial Intelligence Computer Science: An Overview Tenth Edition.
1 Intelligent Autonomous Adaptive Control ( AAC) Method and AAC systems Prof. Alexander ZHDANOV Head of Adaptive control methods Department
Operant Conditioning of Cortical Activity E Fetz, 1969.
Chapter 16. Basal Ganglia Models for Autonomous Behavior Learning in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans.
Autonomous Virtual Humans Tyler Streeter April 15, 2004.
Advanced Robotics Projects for Undergraduate Students Douglas Blank Bryn Mawr College James Marshall Sarah Lawrence College Deepak Kumar Bryn Mawr College.
DARPA ITO/MARS Project Update Vanderbilt University A Software Architecture and Tools for Autonomous Robots that Learn on Mission K. Kawamura, M. Wilkes,
Verve: A General Purpose Open Source Reinforcement Learning Toolkit Tyler Streeter, James Oliver, & Adrian Sannier ASME IDETC & CIE, September 13, 2006.
Curiosity-Driven Exploration with Planning Trajectories Tyler Streeter PhD Student, Human Computer Interaction Iowa State University
Evolving the goal priorities of autonomous agents Adam Campbell* Advisor: Dr. Annie S. Wu* Collaborator: Dr. Randall Shumaker** School of Electrical Engineering.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
Cooperative Q-Learning Lars Blackmore and Steve Block Multi-Agent Reinforcement Learning: Independent vs. Cooperative Agents Tan, M Proceedings of the.
Learning Agents MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way.
Autonomous Virtual Humans Tyler Streeter. Contents Introduction Introduction Implementation Implementation –3D Graphics –Simulated Physics –Neural Networks.
Design and Implementation of General Purpose Reinforcement Learning Agents Tyler Streeter November 17, 2005.
Chapter 6 FLASH CARD CHALLENGE!!!
Chapter 17. The Progress drive hypothesis Course: Robots Learning from Humans Yoon, Bo Sung Department of Economics Seoul National University
Evolutionary Robotics The French Approach Jean-Arcady Meyer Commentator on the growth of the field. Animats: artificial animals anima-materials Coined.
Constraint-Based Motion Planning for Multiple Agents Luv Kohli COMP259 March 5, 2003.
Chapter 6 Behavior Management 6 Behavior Management E. Michael Loovis C H A P T E R.
Done by Fazlun Satya Saradhi. INTRODUCTION The main concept is to use different types of agent models which would help create a better dynamic and adaptive.
Dynamics of Learning & Distributed Adaptation
Eick: Introduction Machine Learning
Cognitive Computing…. Computational Neuroscience
Active Learning using Adaptive Curiosity
Intelligent Adaptive Mobile Robots
Bioagents and Biorobots David Kadleček, Michal Petrus, Pavel Nahodil
What is AI All About? What is AI All About?
Dynamics of Learning & Distributed Adaptation James P
Learning: Operant Conditioning.
Biology, Cognition, and Learning
CSE 473 Introduction to Artificial Intelligence Neural Networks
Emotions cse 574 winter 2004.
Overview of Machine Learning
INTRODUCTION.
Presentation transcript:

Artificial Curiosity Tyler Streeter

What is Curiosity? An intrinsic drive towards “interesting” situations “Many interesting things are unexpected, but not all unexpected things are interesting.” – Juergen Schmidhuber

Curiosity Rewards Humans and many animals learn from reinforcements (rewards/punishment) Examples of primary reinforcements: food, sex, pain… and “interestingness?” Evidence: primate dopamine neurons fire reward signals in novel situations –Kakade, S. & Dayan, P Dopamine: Generalization and Bonuses. Neural Networks, 15,

So What is “Interesting?” “…situations which are neither too predictable nor too unpredictable.” “…the edge of order and chaos in the cognitive dynamics.” –Oudeyer, P. & Kaplan, F Intelligent Adaptive Curiosity: A Source of Self-Development. In Proceedings of the 4th International Workshop on Epigenetic Robotics, Volume 117, Based on previous experience (very subjective!)

Why Artificial Curiosity? Robots and software agents today depend on human teachers; they lack an intrinsic drive to seek new information

Hypothesis Curious agents will learn broader, more robust sets of skills (even if they must develop more slowly)

Examples of Autonomous Self- Development Stoytchev, A., "Autonomous Learning of Tool Affordances by a Robot", In Proceedings of the Twentieth National Conference on Artificial Intelligence (AAAI), Pittsburgh, Pennsylvania, July 9-13, 2005.

Examples of Autonomous Self- Development Oudeyer, P., Kaplan, F.; Hafner, V.V., & Whyte, A The Playground Experiment: Task-Independent Development of a Curious Robot. In Proceedings of the AAAI Spring Symposium Workshop on Developmental Robotics.

How to Model Curiosity Agents must develop predictive models of their environments –“Based on what I’m seeing, what will happen next?” Predictive models enable simulated experiences/imagination

How to Model Curiosity Reward agents in situations that become more predictable over time - S chmidhuber, J Curious Model-Building Control Systems. In Proceedings of the International Joint Conference on Neural Networks, Singapore, Volume 2, Prediction Errors Boring (Too Predictable) Time Prediction Errors Boring (Too Unpredictable) Time Prediction Errors Interesting! Time

Open Source Implementation Verve: an open source implementation of curious software agents, developed by Tyler Streeter

Interactive Demo Tyler Streeter, Curious mobile robot demo, HCI Forum 2006