Chapter 16. Basal Ganglia Models for Autonomous Behavior Learning in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans.

Slides:



Advertisements
Similar presentations
Pre-frontal cortex and Executive Function Squire et al Ch 52.
Advertisements

Learning - Dot Point 2. Part A. Learning and Changes in the Brain – Brain Structures Associated with Learning.
1 Reinforcement Learning Problem Week #3. Figure reproduced from the figure on page 52 in reference [1] 2 Reinforcement Learning Loop state Agent Environment.
AI 授課教師:顏士淨 2013/09/12 1. Part I & Part II 2  Part I Artificial Intelligence 1 Introduction 2 Intelligent Agents Part II Problem Solving 3 Solving Problems.
Reinforcement learning (Chapter 21)
COSC 878 Seminar on Large Scale Statistical Machine Learning 1.
Copyright © 2006 by Allyn and Bacon Chapter 8 The Sensorimotor System How You Do What You Do This multimedia product and its contents are protected under.
Dopamine, Uncertainty and TD Learning CNS 2004 Yael Niv Michael Duff Peter Dayan Gatsby Computational Neuroscience Unit, UCL.
Fundamentals of Cognitive Psychology
CS 182/CogSci110/Ling109 Spring 2008 Reinforcement Learning: Details and Biology 4/3/2008 Srini Narayanan – ICSI and UC Berkeley.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Episodic Control: Singular Recall and Optimal Actions Peter Dayan Nathaniel Daw Máté Lengyel Yael Niv.
Vedrana Vidulin Jožef Stefan Institute, Ljubljana, Slovenia
Neural circuits for bias and sensitivity in decision-making Jan Lauwereyns Associate Professor, Victoria University of Wellington, New Zealand Long-term.
Anthony Correa Stephanie Regan Kasha Geels
LIMBIC SYSTEM NBIO 401 Robinson. Objectives: -1) Be able to describe the major inputs and outputs, function, and the consequences of lesions or electrical.
REINFORCEMENT LEARNING LEARNING TO PERFORM BEST ACTIONS BY REWARDS Tayfun Gürel.
Prediction in Human Presented by: Rezvan Kianifar January 2009.
Reinforcement learning and human behavior Hanan Shteingart and Yonatan Loewenstein MTAT Seminar in Computational Neuroscience Zurab Bzhalava.
Chapter 1. Introduction in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans Jo, HwiYeol Biointelligence Laboratory.
Matt Mehne, USC Brain Project 1 USC Brain Project Matt Mehne – Progress Update – May 15, 2007 Modeling Action Selection in the Parkinsonian Basal Ganglia.
Michael S. Beauchamp, Ph.D. Assistant Professor Department of Neurobiology and Anatomy University of Texas Health Science Center at Houston Houston, TX.
Testing computational models of dopamine and noradrenaline dysfunction in attention deficit/hyperactivity disorder Jaeseung Jeong, Ph.D Department of Bio.
Background The physiology of the cerebral cortex is organized in hierarchical manner. The prefrontal cortex (PFC) constitutes the highest level of the.
CS344 : Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 26- Reinforcement Learning for Robots; Brain Evidence.
Learning Automata based Approach to Model Dialogue Strategy in Spoken Dialogue System: A Performance Evaluation G.Kumaravelan Pondicherry University, Karaikal.
Show Me the Money! Dmitry Kit. Outline Overview Reinforcement Learning Other Topics Conclusions.
The role of the basal ganglia in habit formation Group 4 Youngjin Kang Zhiheng Zhou Baoyu Wang.
CHAPTER 10 Reinforcement Learning Utility Theory.
Bayesian Reinforcement Learning Machine Learning RCC 16 th June 2011.
Reinforcement Learning
Curiosity-Driven Exploration with Planning Trajectories Tyler Streeter PhD Student, Human Computer Interaction Iowa State University
Neural Networks Chapter 7
PHYSIOLOGICAL UNDERPINNINGS OF LANGUAGE, PROBLEM SOLVING, AND REASONING.
INTRODUCTION TO Machine Learning
A View from the Bottom Peter Dayan Gatsby Computational Neuroscience Unit.
CHAPTER 16: Reinforcement Learning. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Introduction Game-playing:
1 Introduction to Reinforcement Learning Freek Stulp.
Figure 18.1 Motor components of the human basal ganglia.
Chapter 2. From Complex Networks to Intelligent Systems in Creating Brain-like Systems, Sendhoff et al. Course: Robots Learning from Humans Baek, Da Som.
Reinforcement learning (Chapter 21)
Reinforcement Learning
Chapter 3. Stochastic Dynamics in the Brain and Probabilistic Decision-Making in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning.
Vedrana Vidulin Jožef Stefan Institute, Ljubljana, Slovenia
Chapter 2. From Complex Networks to Intelligent Systems in Creating Brain-Like Intelligence, Olaf Sporns Course: Robots Learning from Humans Park, John.
Chapter 14. Active Vision for Goal-Oriented Humanoid Robot Walking (2/2) in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from.
A B SAMPLE DELAY CHOICE A B A B 1st Reversal2nd Reversal etc… Trial Time (ms) Conditional Visuomotor Learning Task Asaad, W.F., Rainer, G. and.
Lars Taxén – Activity Modalities An action perspective based on innate coordination capacities Lars Taxén, Linköping University
Chapter 15. Cognitive Adequacy in Brain- Like Intelligence in Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans Cinarel, Ceyda.
Chapter 1. Introduction in Creating Brain-like intelligence, Sendhoff et al. Course: Robots Learning from Humans Bae, Eun-bit Otology Laboratory Seoul.
Chapter 4. Analysis of Brain-Like Structures and Dynamics (2/2) Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans 09/25.
Neural correlates of risk sensitivity An fMRI study of instrumental choice behavior Yael Niv, Jeffrey A. Edlund, Peter Dayan, and John O’Doherty Cohen.
Neurochemistry of executive functions
The Brain. The Big Questions What do different parts of the brain do, and how do they work together?  Information flows through neighborhoods..  Different.
18 Actions, Habits, and the Cortico-Striatal System.
Cuneacerebellar Rostrocerebellar Reticolocerebellar Rubrocerebellar Input pathways to the cerebellum.
Neural Coding of Basic Reward Terms of Animal Learning Theory, Game Theory, Microeconomics and Behavioral Ecology Wolfram Schultz Current Opinion in Neurobiology.
Organization of Neural Structures Involved in the Control of Movement
Reinforcement Learning
Joelle Pineau: General info
Basal ganglia movement modulation
Organization of Neural Structures Involved in the Control of Movement
A C B.
CS4341 Introduction to Artificial Intelligence
Dr. Unnikrishnan P.C. Professor, EEE
Presented by: Rezvan Kianifar January 2009
CHAPTER I. of EVOLUTIONARY ROBOTICS Stefano Nolfi and Dario Floreano
Wallis, JD Helen Wills Neuroscience Institute UC, Berkeley
Circuitry of self-control and its role in reducing addiction
Presentation transcript:

Chapter 16. Basal Ganglia Models for Autonomous Behavior Learning in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans Kim, Jung Ah College of Natural Sciences Interdisciplinary Program in Brain Science Seoul National University

Contents 1.Introduction 1.Related Works – Reinforcement Learning Research – Neuroscience Findings – Open Questions 2.Basal Ganglia Models – Basal Ganglia System Model – Basal Ganglia Spiking Neural Network Model 1.Discussion 2.Conclusion 2

Introduction Research objective : – to develop an autonomous behavior learning system for machines Why? – People will demand that the machine selects the best option by assessing the situation What to investigate: – Learning mechanism underlying behavior selection in the animal basal ganglia (BG)

Introduction Animal Learning Basal Ganglia (BG) Classical Pavlovian conditioning Instrumental conditioning Machine Learning Reinforcement Learning (RL) Temporal difference (TD) learning Phasic activity of dopamine neurons in the BG

Related Works (1) – What is RL? Action a t Reward r t State s t+1 In reinforcement learning: agent interacts with its environment perceptions (state), actions, rewards (repeat) task is to choose actions to maximize rewards complete background knowledge unavailable Reinforcement learning (RL) is concerned with “how an agent ought to take actions in an environment so as to maximize some notion of long-term reward.”

Related Works (2) – BG Cerebral Cortex (frontal, prefrontal, and parietal areas) BG (striatum) Cerebral cortex (motor area) Basal Ganglia(BG) is buried deep within the telencephalon

Related Works (2) – Why BG? Information from the cortex flows through the direct and indirect pathways in parallel The outputs of both pathways ultimately regulate the motor thalamus The direct pathway helps to select certain motor actions while the indirect pathway simultaneously suppresses competing, and inappropriate, motor programs

Related Works Machine learning (Reinforcement learning, RL) Neuroscience (Basal Ganglia, BG) Model-based RL Uses experiences to construct an internal model of state transitions Can solve complex structured tasks Dyna and real-time dynamic programming Model-free RL Uses experiences to directly learn one or two simpler quantities, which can then achieve the optimal behavior without learning a world model Can be used to solve unknown tasks Temporal difference(TD) learning and Q-learning Model-based RL Dorsomedial striatum Prelimbic prefrontal cortex Orbito frontal cortex Medial prefrontal cortex Parts of the amygdala Model-free RL (Neuromodulatory system) Dorsolateral striatum Amygdala These brain areas are interconnected by parallel loops! BG are running model-free RL independently, as well as running model-based RL by receiving modeled, or structured, input from cortex.

Basal Ganglia(BG) Models (I) BG System Model Two types of test environments: MDP (Markov Decision Process) HOMDP (High Order MDP)

Basal Ganglia(BG) Models (I) BG System Model Two types of test environments: 1) MDP (Markov Decision Process) 2) HOMDP (High Order Markov Decision Process)

Basal Ganglia(BG) Models (II) BG Spiking Neural Network Model – Can select and initiate an action for trial and error in the presence of noisy, ambiguous input streams, and then adaptively of actions tune selection probability and timing – Indirect pathway selects an action and the direct pathway initiates the selected action

Discussion The BG system model focuses on the problems associated with relating model-based RL and model- free RL in a single system Open questions: – Architecture – Role of Neuromodulators – Neural Mechanisms of the Timing Perspective: – Associative Interacting Intelligence

Conclusion The BG system model illustrates the effectiveness of internal state representation and internal reward for achieving a goal in shorter trials The BG spiking neural circuit model has the capacity for probabilistic selection of action and also shows that selection probability and execution timing can be modulated