18 Actions, Habits, and the Cortico-Striatal System.

Slides:



Advertisements
Similar presentations
Instrumental Conditioning Also called Operant Conditioning.
Advertisements

Social defeat stress, sensitization, and intravenous cocaine self-administration By Jasmine Yap and Klaus Miczek.
Learning - Dot Point 2. Part A. Learning and Changes in the Brain – Brain Structures Associated with Learning.
PSY402 Theories of Learning Chapter 9, Theories and Applications of Aversive Conditioning.
Lecture 17: Instrumental Conditioning (Associative Structures) Learning, Psychology 5310 Spring, 2015 Professor Delamater.
Learning & Memory Sean Montgomery - TA Behavioral and Cognitive Neuroanatomy
Traditional Learning Theories
Chapter 3 Basic Vehicle Control
PSY 402 Theories of Learning Chapter 7 – Behavior & Its Consequences Instrumental & Operant Learning.
Chapter 8 Instrumental Conditioning: Learning the Consequences of Behavior.
This device tells you how fast you are driving. SPEEDOMETER.
Competitive Place Task
© 2013 by McGraw-Hill Education. This is proprietary material solely for authorized instructor use. Not authorized for sale or distribution in any manner.
LIMBIC SYSTEM NBIO 401 Robinson. Objectives: -1) Be able to describe the major inputs and outputs, function, and the consequences of lesions or electrical.
Chapter 5 Learning. chapter 5 What is Learning? Occurs whenever experience or practice results in a relatively permanent change in behavior.
Box 29A(1) Facial Expressions: Pyramidal and Extrapyramidal Contributions.
CS344 : Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 26- Reinforcement Learning for Robots; Brain Evidence.
Copyright McGraw-Hill, Inc Chapter 5 Learning.
The role of the basal ganglia in habit formation Group 4 Youngjin Kang Zhiheng Zhou Baoyu Wang.
Basic Behavioral Neurobiology of Drug Addiction J. David Jentsch, Ph.D. Assistant Professor of Psychology (Behavioral Neuroscience)
Chapter 16. Basal Ganglia Models for Autonomous Behavior Learning in Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans.
Instrumental/Operant Conditioning. Thorndike’s Puzzle Box.
Module 10 Operant & Cognitive Approaches. Thorndike’s Law of Effect l Behaviors followed by positive consequences are strengthened while behaviors followed.
The Role of the Basal Ganglia in Habit Formation By Henry H. Yin & Barbara J. Knowlton Group 3, Week 10 Alicia Iafonaro Kimberly Villalva Tawni Voyles.
Neurological Disorders Lesson 5.1 What circuit do drugs affect in our brains?
Choice1Choice2Choice3Choice4Choice
PHYSIOLOGICAL UNDERPINNINGS OF LANGUAGE, PROBLEM SOLVING, AND REASONING.
If behavior was dominated in the past by Hull’s S-R reinforcement paradigm, what paradigm is it dominated by today? There is a causal relationship between.
Chapter 5 Learning. Copyright © 1999 by The McGraw-Hill Companies, Inc. 2 Defining Learning Learning –a relatively permanent change in behavior that occurs.
3.1 Instruments, Controls, & Devices. _______________ Tells you the speed you are traveling in both _______________ and _______________. Some have digital!
Chapter 11 Inferences and the Representations of Knowledge in Operant Conditioning.
Unit 1 Review 1. To say that learning has taken place, we must observe a change in a subject’s behavior. What two requirements must this behavioral change.
BASIC VEHICLE CONTROL CHAPTER 3. Basic Vehicle Control When you begin driving, you will need to know the instruments, controls, and devices that you will.
Driver’s Education Chapter 3 Basic Vehicle Control.
Trial and error learning Thorndike’s puzzle box. Trial and error learning This type of learning occurs when an organism attempts to learn by undertaking.
Learning. What is Learning? Acquisition of new knowledge, behavior, skills, values, preferences or understanding.
Edward Thorndike Research and Theory. Born in Williamsburg, Massachusetts (August 31, 1874) Earned BS from Wesleyan University-1895 MA at Harvard University-1897.
Chapter 6 LEARNING. Learning Learning – A process through which experience produces lasting change in behavior or mental processes. Behavioral Learning.
Learning Chapter 5 Presentation:Fajr Harris Presenter:Daniel Rodriguez
What causes cigarette cravings?
Chapter 5 Learning © 2013 by McGraw-Hill Education. This is proprietary material solely for authorized instructor use. Not authorized for sale or distribution.
Chapter 3 Basic Vehicle Control
Prior cocaine exposure disrupts extinction of fear conditioning
Dopamine system: neuroanatomy
Unit 2 Basic Vehicle Control
Module 20 Operant Conditioning.
Do “good” things motivate us?
How Do We Know How the Brain Works?: 6 Essential Pathways
According to the Specification you need to be able to :
Chapter 3 Basic Vehicle Control.
Operant Conditioning Unit 4 - AoS 2 - Learning.
UNIT 4 BRAIN, BEHAVIOUR & EXPERIENCE
Dopamine pathways & antipsychotics
Alexander W. Johnson  Trends in Neurosciences 
Memory Gateway to Learning.
Illegal Drug Use in the U.S.
Conditioning and Learning
Thorndike Interested in studying animal intelligence
Dopamine in Motivational Control: Rewarding, Aversive, and Alerting
Human Learning.
Ronald Keiflin, Patricia H. Janak  Neuron 
Wallis, JD Helen Wills Neuroscience Institute UC, Berkeley
Brain Reward Circuitry
Circuitry of self-control and its role in reducing addiction
PSY 402 Theories of Learning Chapter 7 – Behavior & Its Consequences
lesson 3.3 STARTING, STOPPING, STEERING, AND TARGETING
Learning.
Module 16 Emotion.
Thorndike Interested in studying animal intelligence
Orbitofrontal Cortex as a Cognitive Map of Task Space
Presentation transcript:

18 Actions, Habits, and the Cortico-Striatal System

To appreciate the basic problems addressed in this chapter, you might think back to when you initially learned to drive a car. In order to competently drive you had to learn and coordinate many complex behaviors, such as: Insert the ignition key Turn on the ignition key Release the handbrake Put the car in gear Put your foot on the accelerator Generate just the right amount of gas Adjust the steering wheel to maintain or change the trajectory Make proper fine-grain adjustments to oncoming traffic Apply just the right amount of pressure on the brake to stop Some Behaviors Are Modified by the Outcomes They Produce

Most psychologists use the term instrumental learning or instrumental behavior when referring to the study of how behavior is modified by the outcome it produces. This term recognizes that our behaviors can be viewed as instruments that can change or modify our environments. For example, when you turn the ignition key, the engine starts. The Concept of Instrumental Behavior

The Scientific Study of Instrumental Behavior Began with E.L. Thorndike

The essence of Thorndike’s theory is that outcomes produced by behavior ultimately adapt the animal to the situation by strengthening and weakening existing stimulus–response (S–R) connections. Outcomes that reward behavior strengthen S–R connections; nonrewarding outcomes weaken connections. Thorndike’s Law of Effect

According to E.C. Tolman: Instrumental behaviors are organized around goals and mediated by expectancies. An expectancy is a three-term association (S1–R–S2) that includes a representation of the stimulus situation (S1) that preceded the response, a representation of the response (R), and a representation of the outcome (S2) produced by the response. Tolman’s Cognitive Expectancy Theory

Comparison of the S–R Habit versus Expectancy Representation of the Cat’s Solution to the Puzzle Box

Instrumental Behaviors Can Be Classified as Either Actions or Habits Actions and habits differ on four dimensions.

The reward devaluation strategy centers on changing the value of the outcome after the animal has learned to solve the problem. How Can You Tell If an Instrumental Behavior Is an Action or a Habit? The Reward Devaluation Strategy

The Rat’s Lever-Pressing Response Becomes a Habit with Extensive Training Instrumental behaviors are initially controlled by the action system. This conclusion is based on the fact that after limited training the animals behavior is reduced following reward devaluation. However, with extensive practice the behaviors become habits. This conclusion is based on the fact that after extensive training, reward devaluation has no effect compared to the control condition.

Conceptual Model for Actions and Habits

Actions versus Habits Action and habit systems compete for control. Once an instrumental response is well learned, the two systems can work synergistically to support the same behavior, with the habit system dominating. However, when the contingencies are reversed, the action system attempts to rapidly adjust to the system but the output from the habit system can interfere with adaptations. Example: An American crossing the street in England has to overcome the habit of looking to the left before crossing.

The Acquisition of Instrumental Behaviors Depends on the Striatum Rats with damage to the striatum could not learn that immediately turning left or right to enter the adjacent arm would be rewarded.

The Striatum Integrates Information from Many Brain Regions to Create Instrumental Behaviors

Neural Support for the Action System: Dorsomedial Striatum Supports Actions Rats with damage to the dorsomedial striatum are not sensitive to reward devaluation.

Neural Support for the Action System: The Basolateral Amygdala Attaches Value Outcomes The amygdala is critical for attaching value to a particular outcome. Primates and rodents with damage to this region learn simple discriminations but are insensitive to reward devaluation. Sham Amygdala Con Deval Number of Response/min

The prelimbic region is critical in the acquisition of the associations that support an action. We know this because when this region is damaged prior to training, reward devaluation has no effect, meaning the behavior is a habit. The prelimbic region is not the site in the brain where these associations are stored. We know this because if this region is damaged after training, rats are sensitive to reward devaluation. Neural Support for Actions: Prelimbic Prefrontal Cortex Supports Action Learning but Is Not a Storage Site Prelimbic Prior After Con Deval Number of Response/min

Damage to the dorsolateral striatum prevents the development of habits. Following extensive lever press training, normal rats are insensitive to reward devaluation. However, rats with damage to the dorsolateral striatum are still sensitive to devaluation, indicating that the behavior never became a habit. Neural Support for Habits: Dorsolateral Striatum DLS damage Control DLS lesion Con Deval Number of Response/min

Damage to this region prior to training prevents instrumental behaviors from becoming habits. Rats with such damage remain sensitive to reward devaluation even after extensive training. Moreover, if this region is damaged after rats have be extensively trained and the instrumental response has become a habit, the rats again become sensitive to reward devaluation. All this suggests that with extensive training the infralimbic region comes to suppress the output of the action system. Infralimbic Medial Prefrontal Cortex Suppresses the Action System Infralimbic Prior After Con Deval Number of Response/min

The Contributions of the Prelimbic and Infralimbic Regions of the Prefrontal Cortex to Instrumental Behavior

The Contributions of the Prelimbic and Infralimbic Regions of the Medial Prefrontal Cortex to Instrumental Behavior Summary Prelimbic prefrontal cortex contributes to creating action based behavior. However, its not a storage site for the action pattern because damage to it after the action has been learned does not affect animals sensitive to reward devaluation. Infralimbic prefrontal cortex suppresses the action system. If this region is damaged before training the animal never becomes insensitive to reward devaluation even after extensive training. Moreover after extensive training, when the behavior is normally a habit, if the infralimbic region is damaged it again becomes action—sensitive to reward devaluation.

The Mesolimbic Dopamine System and the Dopamine Theory of Reinforcement (A) The mesolimbic dopamine system. The dopamine neurons are located in the ventral tegmental area (VTA) and their fibers project into the nucleus accumbens. (B) The dopamine theory of reinforcement. An outcome–reward has two functions: (1) It generates a representation (O), and (2) it activates dopamine neurons in the VTA that release dopamine. This strengthens synaptic connections between the representations of the stimulus (S) complex and the response (R) and perhaps between the representations of the response and the outcome.

The Incentive Salience Hypothesis of How Rewards Influence Instrumental Behavior The incentive salience hypothesis assumes that the reward turns on dopamine neurons in the VTA. In the normal sequence of events that establish instrumental behavior, the stimulus not only can associate with the response, it also can become associated with the incentive properties of the outcome. The stimulus itself can Then elicit strong urges or wants that lead the individual to seek out the reward.

Neural Support for Action and Habit Systems: Summary Instrumental Behavior Action System Dorsomedial striatum Basolateral amygdala Prelimbic prefrontal cortex Habit System Dorsolateral striatum Infralimbic prefrontal cortex Mesolimbic dopamine system