Lectures 15 & 16: Instrumental Conditioning (Schedules of Reinforcement) Learning, Psychology 5310 Spring, 2015 Professor Delamater.

Slides:



Advertisements
Similar presentations
Schedules of reinforcement
Advertisements

Chapter 6 – Schedules or Reinforcement and Choice Behavior
Steven I. Dworkin, Ph.D. 1 Choice and Matching Chapter 10.
Schedules of Reinforcement: Continuous reinforcement: – Reinforce every single time the animal performs the response – Use for teaching the animal the.
The Matching Law Richard J. Herrnstein. Reinforcement schedule Fixed-Ratio (FR) : the first response made after a given number of responses is reinforced.
Chapter 8 Learning.  Learning  relatively permanent change in an organism’s behavior due to experience.
Schedules of Reinforcement There are several alternate ways to arrange the delivery of reinforcement A. Continuous reinforcement (CRF), in which every.
Developing Behavioral Persistence Through the Use of Intermittent Reinforcement Chapter 6.
Quiz #3 Last class, we talked about 6 techniques for self- control. Name and briefly describe 2 of those techniques. 1.
Sniffy the Virtual Rat Psych 210 Winter Lab Assignment.
Copyright © 2011 Pearson Education, Inc. All rights reserved. Developing Behavioral Persistence Through the Use of Intermittent Reinforcement Chapter 6.
Operant Conditioning. Shaping shaping = successive approximations toward a goal a process whereby reinforcements are given for behavior directed toward.
Learning Operant Conditioning.  Operant Behavior  operates (acts) on environment  produces consequences  Respondent Behavior  occurs as an automatic.
Myers EXPLORING PSYCHOLOGY (6th Edition in Modules) Module 19 Operant Conditioning James A. McCubbin, PhD Clemson University Worth Publishers.
Chapter 8 Operant Conditioning.  Operant Conditioning  type of learning in which behavior is strengthened if followed by reinforcement or diminished.
 JqjlrHA JqjlrHA.
More Instrumental (Operant) Conditioning. B.F. Skinner Coined the term ‘Operant conditioning’ Coined the term ‘Operant conditioning’ The animal operates.
PSY402 Theories of Learning Chapter 4 (Cont.) Schedules of Reinforcement.
Schedules of Reinforcement Lecture 14. Schedules of RFT n Frequency of RFT after response is important n Continuous RFT l RFT after each response l Fast.
PSY 402 Theories of Learning Chapter 7 – Behavior & Its Consequences Instrumental & Operant Learning.
PSY402 Theories of Learning Chapter 7 – Theories and Applications of Appetitive Conditioning.
PSY 402 Theories of Learning Chapter 7 – Behavior & Its Consequences Instrumental & Operant Learning.
OPERANT CONDITIONING DEF: a form of learning in which responses come to be controlled by their consequences.
Learning the Consequences of Behavior
Chapter 7 Operant Conditioning:
Week 5: Increasing Behavior
Ratio Schedules Focus on the number of responses required before reinforcement is given.
Psychology of Learning EXP4404 Chapter 6: Schedules of Reinforcement Dr. Steve.
Chapter 9 Adjusting to Schedules of Partial Reinforcement.
Operant Conditioning: Schedules and Theories Of Reinforcement.
Chapter 6 Operant Conditioning Schedules. Schedule of Reinforcement Appetitive outcome --> reinforcement –As a “shorthand” we call the appetitive outcome.
Ninth Edition 5 Burrhus Frederic Skinner.
Operant Conditioning: Schedules and Theories of Reinforcement
Organizational Behavior Types of Intermittently Reinforcing Behavior.
Reinforcement Consequences that strengthen responses.
Overview of the Day Learning - Part 2 Operant conditioning Observational learning.
Learning (Part II) 7-9% of AP Exam Classical Conditioning UCS + UCR + N, etc… Acquisition Extinction Biological Predisposition Pavlov Watson Operant Conditioning.
Chapter 3 Learning (II) Operant (Instrumental) Conditioning.
Classical Conditioning
Chapter 13: Schedules of Reinforcement
Chapter 6 Developing Behavioral Persistence Through the Use of Intermittent Reinforcement.
Operant Conditioning E.L. Thorndike and B.F. Skinner.
Motivation Motivation: The desire to exert effort to perform a particular task. Motivation: The desire to exert effort to perform a particular task. Some.
PSY402 Theories of Learning Chapter 6 – Appetitive Conditioning.
Copyright McGraw-Hill, Inc Chapter 5 Learning.
Schedules of Reinforcement 11/11/11. The consequence provides something ($, a spanking…) The consequence takes something away (removes headache, timeout)
Schedules of Reinforcement and Choice. Simple Schedules Ratio Interval Fixed Variable.
Schedules of Reinforcement CH 17,18,19. Divers of Nassau Diving for coins Success does not follow every attempt Success means reinforcement.
Schedules of Reinforcement Thomas G. Bowers, Ph.D.
Schedules of Reinforcement or Punishment: Ratio Schedules
Schedules of reinforcement
What are the types of predetermined plans for timing the delivery of a reinforcement?
Schedules of Reinforcement
Working Hypothesis: If created by conditioning, could be eliminated using principles of conditioning Behavior Therapy: treatment based on environmental.
1 Quiz Question: In what way are fixed-ratio (FR) and variable-ratio (VR) reinforcement schedules: (a) similar? (b) different?
Reinforcement Schedules 1.Continuous Reinforcement: Reinforces the desired response each time it occurs. 2.Partial Reinforcement: Reinforces a response.
Schedules of Reinforcement
Schedules and more Schedules
Factors Affecting Performance on Reinforcement Schedules
Choice Behavior One.
Operant Conditioning A form of learning in which behavior becomes more or less probable depending on its consequences Associated with B.F. Skinner.
Choice Behavior Two.
Schedules of Reinforcement
Learning: Operant Conditioning.
Operant Conditioning, Continued
Schedules of Reinforcement
Schedules of Reinforcement
9.2 Operant Conditioning “Everything we do and are is determined by our history of rewards and punishments.” –BF Skinner Operant Conditioning: learning.
Learning and Motivation
Presentation transcript:

Lectures 15 & 16: Instrumental Conditioning (Schedules of Reinforcement) Learning, Psychology 5310 Spring, 2015 Professor Delamater

Instrumental Learning: Simple Schedules of Reinforcement 1.Ratio Schedules (Fixed or Variable) 2.Interval Schedules (Fixed or Variable) A “schedule of reinforcement” defines the rules by which reward will be given to an organism for responding. Rewards can be made to depend upon a set # of responses. This is characteristic of a “ratio” schedule. They can also be made to depend upon a set amount of elapsed time. This is characteristic of an interval schedule. Further, the set # or set amount of time can be fixed or variable. For example: FR 10, VR 10, FI 60s, VI 60s Life is learning to adapt to a bunch of reinforcement schedules…

Instrumental Learning: Simple Schedules of Reinforcement 1.Examples of a Ratio Schedules: CRF (every response is rewarded) PRF (partial reinforcement – only some Rs rewarded) FR 1 = CRF FR 10: 10 responses required to get a reward (salesman working on commission) VR 10: on average 10 responses required (playing a slot machine…) 2.Examples of Interval Schedules FI 60s: 1 st response after 60 s is rewarded (Studying for exams throughout the term) VI 60s: 1 st response after X s is rewarded (where x=60s on average) (studying with pop quizzes)

Instrumental Learning: Simple Schedules of Reinforcement Different schedules produce different response patterns Response patterns can be read from a “cumulative recorder” Cumulative responses is indicated over time. The slope of the line is a measure of response rate – steeper slope, faster rate.

Instrumental Learning: Simple Schedules of Reinforcement Different schedules produce different response patterns FR – Pause and Run pattern VR – Steady rate (fairly quick, though not as quick as run phase of FR) FI – Scallop pattern (slow initially, increase as interval approaches) VI – Steady rate (slower than VR)

Instrumental Learning: Simple Schedules of Reinforcement Ratio schedules promote faster rates of responding than interval schedules With overall reinforcement rates held constant, VR produces faster responding than VI.

Instrumental Learning: Simple Schedules of Reinforcement Ratio schedules promote faster rates of responding than interval schedules. WHY? Feedback functions for ratio and interval schedules differ: Reinforcement rate increases with Response rate on Ratio, but NOT on Interval Schedules

Instrumental Learning: Simple Schedules of Reinforcement Ratio schedules promote faster rates of responding than interval schedules. WHY? 1. Feedback functions for ratio and interval schedules differ: Reinforcement rate increases with Response rate on Ratio, but NOT on Interval Schedules 2.Also, Ratio schedules reinforce short IRTs (inter-response times) whereas Interval schedules reinforce longer IRTs.

Instrumental Learning: Choice Concurrent Schedules of Reinforcement 1.Two simple schedules are in effect simultaneously e.g., Conc VI 60s FR 10

Instrumental Learning: Choice Herrnstein’s Matching Law - Relative Rate of Responding matches the Relative Rate of Reinforcement B A r A (B A + B B )(r A + r B ) =

Instrumental Learning: Choice Herrnstein’s Matching Law - Relative Rate of Responding matches the Relative Rate of Reinforcement B A r A (B A + B B )(r A + r B ) = This means that choice of any one behavior Reflects the degree to which OTHER behaviors Can be reinforced in the situation. This could be Related to problem behaviors like drug abuse, unprotected sex, criminality, etc.

Instrumental Learning: Choice Herrnstein’s Matching Law – Why does matching occur? 1.Molecular (or Momentary) Maximizing Local probability of reinforcement directs behavior. Animal chooses the response option that has the highest payoff probability at that moment in time.

Instrumental Learning: Choice Herrnstein’s Matching Law – Why does matching occur? 2. Melioration Theory Local rates of reinforcement direct behavior. Animal chooses the response option that has the highest experienced reinforcement rate, and then keeps adjusting until the same experienced rates occur on each side.

Instrumental Learning: Choice Herrnstein’s Matching Law – Why does matching occur? 2. Melioration Theory Local rates of reinforcement direct behavior. Animal chooses the response option that has the highest experienced reinforcement rate, and then keeps adjusting until the same experienced rates occur on each side.

Instrumental Learning: Self Control Concurrent-Chain Schedules Animal makes a “commitment” response to gain access to a simple schedule of reinforcement. Choice Link leads to Terminal Link of the schedule Response choice in the initial link can be studied. In this example, does the bird prefer working for reward after a Variable or Fixed interval? Answer: They prefer the variable.

Instrumental Learning: Self Control Concurrent-Chain Schedules This task can be used to study “Self Control” One option leads to small immediate Reward, while the other option leads to a large but delayed reward. Is the bird Impulsive or Thoughtful??? Answer: It depends on the Delayed Discounting Function.

Instrumental Learning: Self Control Concurrent-Chain Schedules Delayed Discounting Function: The subjective value of the reward decreases as a function of waiting time. The relative values of the two options at any moment (T 1, T 2 ) determines choice.

Instrumental Learning: Self Control Concurrent-Chain Schedules Interestingly, how well you are able to display self-control as a child predicts how healthy you become as an adult. Also, an organism’s “delay discounting function” may also predict their proclivity towards developing a substance abuse problem (as revealed in animal studies).