Executive Functions and Control of Processing PDP Class March 2, 2011.

Slides:



Advertisements
Similar presentations
Lecture 18: Temporal-Difference Learning
Advertisements

Introduction to Training and Learning in Neural Networks n CS/PY 399 Lab Presentation # 4 n February 1, 2001 n Mount Union College.
Neural Basis of Cognition Lecture 6 Learning and Memory, Part II.
Pre-frontal cortex and Executive Function Squire et al Ch 52.
Tuomas Sandholm Carnegie Mellon University Computer Science Department
1.Exams due 9am 16 th. (grades due 10am 19 th ) 2.Describe the organization of visual signals in extra-striate visual cortex and the specialization of.
Learning in Recurrent Networks Psychology 209 February 25, 2013.
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Kostas Kontogiannis E&CE
Scatterplots, Association, and Correlation 60 min.
Some concepts from Cognitive Psychology to review: Shadowing Visual Search Cue-target Paradigm Hint: you’ll find these in Chapter 12.
Midterm 1 Wednesday next week!. Synthesize the Big Picture Understanding Brain-wide neural circuits Extracranial electrophysiology EEG/MEG Metabolic Imaging.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Stochastic Neural Networks, Optimal Perceptual Interpretation, and the Stochastic Interactive Activation Model PDP Class January 15, 2010.
Neural Networks Marco Loog.
COGNITIVE NEUROSCIENCE
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
Abstract Cognitive control processes reduce the effects of irrelevant or misleading information on performance. We report a study suggesting that effective.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Stochastic Interactive Activation and Interactive Activation in the Brain PDP Class January 20, 2010.
How facilitation influences an attractor model of decision making Larissa Albantakis.
CS 4700: Foundations of Artificial Intelligence
CS Reinforcement Learning1 Reinforcement Learning Variation on Supervised Learning Exact target outputs are not given Some variation of reward is.
Stimulus Control of Operant Behavior Discrimination Generalization Generalization Gradients Peak Shift Concepts Overview of stimulus control of operant.
Organic Brain Syndromes in the Developmentally Disabled A new way of conceptualizing Dysfunction and Cognitive Restructuring Dr. Jay Rao M.B.,B.S., D.P.M.,
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Artificial Neural Network Unsupervised Learning
Attention Part 2. Early Selection Model (Broadbent, 1958) inputdetectionrecognition FI L T E R Only information that passed the filter received further.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Hebbian Coincidence Learning
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 4, 2013.
ARTIFICIAL NEURAL NETWORKS. Overview EdGeneral concepts Areej:Learning and Training Wesley:Limitations and optimization of ANNs Cora:Applications and.
Attention Loads program into working memory (more about that later) Vast amount of perceptual information available at one Moment How much can be retained.
Subject wearing a VR helmet immersed in the virtual environment on the right, with obstacles and targets. Subjects follow the path, avoid the obstacels,
Modelling Language Evolution Lecture 1: Introduction to Learning Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
Control of Attention in schizophrenia 1.Advance understanding of schizophrenia. Move from description of deficits to explanation of specific mechanisms.
Chapter 50 The Prefrontal Cortex and Executive Brain Functions Copyright © 2014 Elsevier Inc. All rights reserved.
Introduction to Neural Networks and Example Applications in HCI Nick Gentile.
Metacognition Neil H. Schwartz Psych 605. Three Terms: Organizing the Concept Exogenous Constructivism Endogenous Constructivism Spence Piaget Dialectical.
The effects of working memory load on negative priming in an N-back task Ewald Neumann Brain-Inspired Cognitive Systems (BICS) July, 2010.
Pattern Classification of Attentional Control States S. G. Robison, D. N. Osherson, K. A. Norman, & J. D. Cohen Dept. of Psychology, Princeton University,
Over-Trained Network Node Removal and Neurotransmitter-Inspired Artificial Neural Networks By: Kyle Wray.
James L. McClelland Stanford University
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Principled Probabilistic Inference and Interactive Activation Psych209 January 25, 2013.
The Emergent Structure of Semantic Knowledge
A review of select models by Dehaene & Changeux and the implications for future work By Robert Schuler June 5, /11/2007.
Chapter 6 Neural Network.
Network Models (2) LECTURE 7. I.Introduction − Basic concepts of neural networks II.Realistic neural networks − Homogeneous excitatory and inhibitory.
High-level attention Attention in complex tasks Central executive function Automaticity Attention deficits.
Does the brain compute confidence estimates about decisions?
Date of download: 6/28/2016 Copyright © 2016 American Medical Association. All rights reserved. From: Teamwork Matters: Coordinated Neuronal Activity in.
Dynamics of Reward Bias Effects in Perceptual Decision Making Jay McClelland & Juan Gao Building on: Newsome and Rorie Holmes and Feng Usher and McClelland.
NEURONAL NETWORKS AND CONNECTIONIST (PDP) MODELS Thorndike’s “Law of Effect” (1920’s) –Reward strengthens connections for operant response Hebb’s “reverberatory.
Machine Learning Supervised Learning Classification and Regression
What does the Stroop effect tell us about perception?
Artificial Neural Networks
Learning in Neural Networks
Dynamical Models of Decision Making Optimality, human performance, and principles of neural information processing Jay McClelland Department of Psychology.
Simple learning in connectionist networks
Dopamine Does Double Duty in Motivating Cognitive Effort
Nils Kolling, Marco Wittmann, Matthew F.S. Rushworth  Neuron 
Volume 40, Issue 6, Pages (December 2003)
Dynamic Coding for Cognitive Control in Prefrontal Cortex
Uma R. Karmarkar, Dean V. Buonomano  Neuron 
Simple learning in connectionist networks
Stephen V. David, Benjamin Y. Hayden, James A. Mazer, Jack L. Gallant 
Presentation transcript:

Executive Functions and Control of Processing PDP Class March 2, 2011

The Stroop Task(s): Read Word RED GREEN BLUE GREEN BLUE RED BLUE RED GREEN BLUE GREEN

The Stroop Task: Name Ink Color BLUE RED BLUE RED GREEN BLUE GREEN RED GREEN BLUE GREEN

Reaction Time Data from a Stroop Experiment Color naming Word reading

Alternative Views of the Role of Control and Attention Two kinds of processing: Controlled processing and automatic processing –One slow and effortful, the other fast and effort-free One kind of processing, varying in ‘pathway strength’ –When strength is low, control is necessary to allow the pathway to control behavior –When strength is high, control is necessary to prevent the pathway from controlling behavior

A Model of Control of Automatic Processing in the Stroop Task (Cohen, Dunbar, and McClelland, 1990) Model employs a feed-forward, non-linear activation process so that color and word input can activate response units. –net i (t) =   a j (t)w ij (t) + (1-)net i (t-1) –a i (t) = logistic(net i (t)) Activation builds up over time (rate controlled by  Speed and automaticity of processing depends on strengths of connections. –Connections are stronger for word than for color, so that the word tends to dominate activation of output units. Activation coming from ‘task demand’ units provides an additional source of input that can regulate the ‘strength’ of each pathway. –This instantiates the ‘Biased Processing’ model of control of behavior and attention. The effect is graded so that strong pathways tend to influence performance even when it would be best if their influence could be prevented.

The Cohen et al Stroop model (fully ‘feed-forward’)

How Task Biases Processing in the Stroop Model Input from task demand units sets the baseline net input for the hidden units. Stimulus effects add or subtract from the baseline. The same size stimulus effect has a bigger effect on activation of units when the baseline is near the middle of the range than when it is down near the bottom. This allows task to amplify task-relevant input to the output units, relative to task-irrelevant inputs.

How the model accounts for key features of the data Left panel below shows how asymptotic activation of response units tends to be more affected by inhibition (as in conflict trials) than excitation (as in congruent trials). Right panel shows how this translates into more slowing of performance (interference) on conflict trials than speeding of performance (facilitation) on congruent trials. < response threshold Net input and activation at asymptote Build up of activation over time Time to reach threshold for different conditions

Effect of practice using color words as names for arbitrary shapes on naming the shape and the shape’s color

AX-CPT: Experimental and Simulation Results First-episode schizophrenics show a deficit only at a long delay, while multi-episode schizophrenics show a deficit at both the short and long delay. Must we postulate two different deficits? Braver et al suggest maybe not, since different degrees of gain reduction can account for each of the two patterns of effects. The performance measure represents the degree to which responding at the time of probe presentation (X) is modulated by prior context (A vs. non-A).

Take-Home Message from the work of Cohen et al. In these models, deficits that look like failures to inhibit pre-potent responses and deficits that look like failures of working memory both arise as a consequence of reduced activation of units responsible for maintained activation of task or context information in a form that allows it to exert control over processing. Thus, the authors suggest, the crucial role of the PFC is to provide the mechanism that maintains this sort of information. The exertion of control requires connections from PFC so that specific content maintained in PFC exerts control on appropriate specific neurons elsewhere.

Evidence for a hierarchy of control Velanova et al (J. Neurosci, 2003) contrasted low vs. high control (well-studied vs. once-studied) retrieval conditions. Upper panel shows left Lateral PFC region showing greater transient activity in the high control condition. Lower panel shows right fronto- polar region showing sustained activity in the high control condition (left) but not the low-control condition (right). From this they argue for a hierarchy of control, with fronto-polar cortex maintaining overall task set and Lateral PFC playing a transient role in controlling memory search on a trial-by trial basis. Ideas about the role of particular PFC regions have continued to grow more and more complex…

But what’s controlling the controller (that’s controlling the controller)? Is there an infinite regress problem in cognitive control? One way to break out of it is with a recurrent network. Learned connection weights in the network can allow an input to set up a state that persists until another input comes in that sets the network in a different state. A model of performance in a working memory task is shown on the next slide.

Working Memory Task Studied by Pelegrino Et Al and Modeled by Moody Et al

Network is trained to ‘load’ x-y value when presented with ‘gate’, continue to output x and y while viewing test distractors, then produce a ‘match’ response when the test stimulus matches the sample. Moody et al’s Network

A storage cell: - firing starts and ends with match - activation is unaffected by distractors

How do we usually maintain information? While Moody et al treat the maintenance of information as a function solely of pre-frontal cortex, others (esp. Fuster) have argued that maintenance is generally a function of a multi- regional maintained activation state in which the PFC plays a role but so does sustained activity in posterior cortex. This of course is more consistent with a PDP approach.

Control as an Emergent Process: Botvinick and Plaut (2004) Model of Action, Action Slips, and Action Disorganization Syndrome Model is a simple recurrent network trained to make coffee or tea (with cream and/or sugar). Network is trained on valid task sequences. Input at the start of a sequence can specify the task (coffee/tea; cream/sugar), but goes away; network must maintain this. At test, it is given a starting input, then its inputs are updated given its actions. Slips and A.D.S. are simulated with the addition of different amounts of noise. The net can learn to do the tasks perfectly. Slips have characteristics of human behavior –Right action with wrong object –Occur at boundaries between sub-tasks –Sensitive to frequency Performance degrades gracefully with more and more noise. –Omissions become more prevalent at high noise levels, as in patients.

Reward Based Learning Thorndike’s ‘Law of Effect’ –Responses the lead to a pleasing outcome are strengthened; those that lead to an annoying outcome are weakened Neural networks can be trained using reinforcement signals rather than explicit ‘do this’ targets, e.g. –Reward modulated Hebbian learning: w ij = Rf(a i )g(a j ) –Or using a forward model –We will explore one approach that addresses what to do when rewards do not immediately follow actions Reward signals in the brain appear to be based on ‘unexpected reward’ –This finding is consistent with the model we will explore

T-D learning Example: Predicting the weather on Friday from the weather today As I get closer to Friday, my predictions get better, until I get it right based on the outcome itself (V k+1 = z) The sum of the differences between my daily predictions add up to the difference between my prediction today and the outcome. I learn how I should have adjusted my prediction at t based on the observation I obtain at t+1. If I just have running predictions and if I believe that proximal valuations should play a stronger role than last valuations, I can use TD(), which discounts past estimates. This gives rise to the concept of the eligibility trace of past prediction error gradients subject to correction based on the observation at t+1

Why learning based on estimated future reward can be better than learning based on the final outcome

Reinforcement learning Discounted returns

Choosing Outcomes Based on Expected Reward (SARSA, Q-learning) Action selection policies: –Greedy (-greedy) –Softmax ()

TDBP Uses a BP network to learn predicted Q values Tessauro (1992, 2002) created TD-gammon using this method, still (?) the world champion. Input: Board configuration, role of dice All possible next moves considered one-by-one Choose next action based on predicted value Discovered better moves for some situations, so human players have learned from TD-gammon

Trash Grid Model Thanks to –Clayton Melina (Stanford SS / CS co-term ‘11) –Steven Hansen (CMU CS ‘12)