Studies on Goal-Directed Feature Learning Cornelius Weber, FIAS presented at: “Machine Learning Approaches to Representational Learning and Recognition.

Slides:



Advertisements
Similar presentations
Object recognition and scene “understanding”
Advertisements

Template design only ©copyright 2008 Ohio UniversityMedia Production Spring Quarter  A hierarchical neural network structure for text learning.
Psych 216: Movement Attention. What is attention? There is too much information available in the world to process it all. Demonstration: change-detection.
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
A saliency map model explains the effects of random variations along irrelevant dimensions in texture segmentation and visual search Li Zhaoping, University.
Image classification by sparse coding.
Learning Convolutional Feature Hierarchies for Visual Recognition
Wake-Sleep algorithm for Representational Learning
Visual Pathways W. W. Norton Primary cortex maintains distinct pathways – functional segregation M and P pathways synapse in different layers Ascending.
On Linking Reinforcement Learning with Unsupervised Learning Cornelius Weber FIGSS talk, FIAS, 20 th April 2009.
Journal club Marian Tsanov Reinforcement Learning.
Neural Networks Basic concepts ArchitectureOperation.
CSE 153 Cognitive ModelingChapter 3 Representations and Network computations In this chapter, we cover: –A bit about cortical architecture –Possible representational.
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
Self-Organizing Hierarchical Neural Network
Un Supervised Learning & Self Organizing Maps Learning From Examples
On Linking Reinforcement Learning with Unsupervised Learning Cornelius Weber, FIAS presented at Honda HRI, Offenbach, 17 th March 2009.
Project funded by the Future and Emerging Technologies arm of the IST Programme FET-Open scheme Neural Robot Control Cornelius Weber Hybrid Intelligent.
December 1, 2009Introduction to Cognitive Science Lecture 22: Neural Models of Mental Processes 1 Some YouTube movies: The Neocognitron Part I:
Chapter Seven The Network Approach: Mind as a Web.
From Exploration to Planning Cornelius Weber and Jochen Triesch Frankfurt Institute for Advanced Studies Goethe University Frankfurt, Germany 18 th International.
Basic Processes in Visual Perception
Neural Networks Lab 5. What Is Neural Networks? Neural networks are composed of simple elements( Neurons) operating in parallel. Neural networks are composed.
Goal-Directed Feature and Memory Learning Cornelius Weber Frankfurt Institute for Advanced Studies (FIAS) Sheffield, 3 rd November 2009 Collaborators:
November 24, 2009Introduction to Cognitive Science Lecture 21: Self-Organizing Maps 1 Self-Organizing Maps (Kohonen Maps) In the BPN, we used supervised.
Outline 1) Goal-Directed Feature Learning (Weber & Triesch, IJCNN 2009)‏ Task 4.1 Visual processing based on feature abstraction 2) Emergence of Disparity.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Lecture 09 Clustering-based Learning
Introduction to machine learning
Unsupervised learning
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
2 2  Background  Vision in Human Brain  Efficient Coding Theory  Motivation  Natural Pictures  Methodology  Statistical Characteristics  Models.
Varieties of Helmholtz Machine Peter Dayan and Geoffrey E. Hinton, Neural Networks, Vol. 9, No. 8, pp , 1996.
An informal description of artificial neural networks John MacCormick.
1 CSC 8520 Spring Paula Matuszek Kinds of Machine Learning Machine learning techniques can be grouped into several categories, in several ways: –What.
1 ECE 517 Reinforcement Learning in Artificial Intelligence Lecture 21: Deep Machine Learning Dr. Itamar Arel College of Engineering Department of Electrical.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
The architecture of the visual system: What is the grand design? April 12, 2010.
The Function of Synchrony Marieke Rohde Reading Group DyStURB (Dynamical Structures to Understand Real Brains)
Human vision Jitendra Malik U.C. Berkeley. Visual Areas.
An Unsupervised Connectionist Model of Rule Emergence in Category Learning Rosemary Cowell & Robert French LEAD-CNRS, Dijon, France EC FP6 NEST Grant.
Real-Time Mapping Systems for Routine and Emergency Monitoring Defining Boundaries between Fairy Tales and Reality A. Brenning (1) and G. Dubois (2) (1)
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Image Segmentation by Complex-Valued Units Cornelius Weber Hybrid Intelligent Systems School of Computing and Technology University of Sunderland Presented.
Copyright Paula Matuszek Kinds of Machine Learning.
SIGNATURE RECOGNITION SYSTEM Group Number:10 Group Members: Richa Goyal(y08uc103) Rashmi Singhal(y08uc102)
Chapter 6 Neural Network.
IJCNN, July 27, 2004 Extending SpikeProp Benjamin Schrauwen Jan Van Campenhout Ghent University Belgium.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
National Taiwan Normal A System to Detect Complex Motion of Nearby Vehicles on Freeways C. Y. Fang Department of Information.
A Neurodynamical Cortical Model of Visual Attention and Invariant Object Recognition Gustavo Deco Edmund T. Rolls Vision Research, 2004.
Machine Learning 12. Local Models.
Goal-Directed Feature and Memory Learning
Self organizing networks
CSE P573 Applications of Artificial Intelligence Neural Networks
Brain States: Top-Down Influences in Sensory Processing
Crowding by a single bar
Continous-Action Q-Learning
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
CSE 573 Introduction to Artificial Intelligence Neural Networks
Sensorimotor Learning and the Development of Position Invariance
Neurocognitive Architecture of Working Memory
Brain States: Top-Down Influences in Sensory Processing
Heterogeneous convolutional neural networks for visual recognition
The Network Approach: Mind as a Web
Volume 27, Issue 2, Pages (August 2000)
November 8, 2010 Dr. Itamar Arel College of Engineering
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Studies on Goal-Directed Feature Learning Cornelius Weber, FIAS presented at: “Machine Learning Approaches to Representational Learning and Recognition in Vision” Workshop at the Frankfurt Institute for Advanced Studies (FIAS), November 27-28, 2008

for taking action, we need only the relevant features x y z

models’ background & overview: - unsupervised feature learning models are enslaved by bottom-up input - reward-modulated activity leads to input selection Nakahara, Neur Comp 14, (2002) - reward-modulated STDP Izhikevich, Cereb Cortex 17, (2007), Florian, Neur Comp 19/6, (2007); Farries & Fairhall, Neurophysiol 98, (2007);... - RL models learn partitioning of input space e.g. McCallum, PhD Thesis, Rochester, NY, USA (1996)‏ - reward-modulated Hebb Triesch, Neur Comp 19, (2007), Roelfsema & Ooyen, Neur Comp 17, (2005); Franz & Triesch, ICDL (2007)‏ (model 3 presented here, extending to delayed reward)‏ - feature-pruning models learn all features but forget the irrelevant ones (models 1 & 2 presented here)‏

sensory input reward action purely sensory data, in which one feature type is linked to reward the action is not controlled by the network

model 1: obtaining the relevant features 1) build a feature detecting model 2) learn associations between features 3) register the average features’ reward 4) spread value along associative connections 5) check whether actions in-/decrease value 6) remove features where action doesn’t matter irrelevant relevant

Földiák, Biol Cybern 64, (1990) → homogeneous activity distr. features thresholds lateral weights (decorrelation)‏ selected features associative weights action effect Weber & Triesch, Proc ICANN, (2008); Witkowski, Adap Behav, 15(1), (2007); Toussaint, Proc NIPS, (2003); Weber, Proc ICANN, (2001)‏ → relevant features indentified

sensory inputreward motor-sensory data (again, one feature type is linked to reward)‏ the network selects the action (to get reward)‏ irrelevant subspace relevant subspace

model 2: removing the irrelevant inputs 1) initialize feature detecting model (but continue learning)‏ 2) perform actor-critic RL, taking the features’ outputs as state representation - works despite irrelevant features - challenge: relevant features will occur at different frequencies - nevertheless, features may remain stable 3) observe the critic: puts negative value on irrelevant features after long training 4) modulate (multiply) learning by critic’s value frequency value

Lücke & Bouecke, Proc ICANN, 31-7 (2005) features critic value action weights → relevant subspace discovered

model 3: learning only the relevant inputs 1) top level: reinforcement learning model (SARSA)‏ 2) lower level: feature learning model (SOM / K-means)‏ 3) modulate learning by δ, in both layers RL weights feature weights input action

model 3: SARSA with SOM-like activation and update

relevant subspace RL action weights subspace coverage feature weights

RL action weights feature weights input reward 2 actions (not shown)‏ data learning ‘long bars’ data

RL action weights feature weights input data: bars controlled by actions ‘up’, ‘down’, ‘left’, ‘right’ learning the ‘short bars’ data reward action

short bars in 12x12 average # of steps to goal: 11

cortex striatum GPi (output of basal ganglia)‏ biological interpretation - no direct feedback from striatum to cortex - convergent mapping → little receptive field overlap, consistent with subspace discovery feature/subspace detection action selection

Discussion - models 1 and 2 learn all features and identify the relevent features - either requires homogeneous feature distribution (model 1)‏ - or can do only subspace- (no real feature) detection (model 2)‏ - model 3 is very simple: SARSA on SOM with δ-feedback - learns only the relevant subspace or features in the first place - link between unsupervised- and reinforcement learning Sponsors Bernstein Focus Neurotechnology EU project “IM-CLeVeR” call FP7-ICT Frankfurt Institute for Advanced Studies FIAS

early learninglate learning Jog et al, Science, 286, (1999)‏ relevant features change during learning units in the basal ganglia are active at the junction during early task acquisition but not at a later stage T - maze decision task (rat)‏

evidence for reward/action modulated learning in the visual system Shuler & Bear, "Reward timing in the primary visual cortex", Science, 311, (2006)‏ Schoups et al. "Practising orientation identification improves orientation coding in V1 neurons" Nature, 412, (2001)‏