Schedules of Reinforcement CH 17,18,19. Divers of Nassau Diving for coins Success does not follow every attempt Success means reinforcement.

Slides:



Advertisements
Similar presentations
Chapter 10 Maintaining Behavior Changes. Relapses in Behavior behavior can regress after goals have been attained a relapse is an extended return to original.
Advertisements

Mean = = 83%
The Matching Law Richard J. Herrnstein. Reinforcement schedule Fixed-Ratio (FR) : the first response made after a given number of responses is reinforced.
Schedules of Reinforcement There are several alternate ways to arrange the delivery of reinforcement A. Continuous reinforcement (CRF), in which every.
Developing Behavioral Persistence Through the Use of Intermittent Reinforcement Chapter 6.
Shaping Chapter 8. Simple Reinforcement Before: Andrew has no gum Behavior Initial: Andrew moves his lips After: Andrew receives gum.
Quiz #3 Last class, we talked about 6 techniques for self- control. Name and briefly describe 2 of those techniques. 1.
Sniffy the Virtual Rat Psych 210 Winter Lab Assignment.
Operant or Instrumental Conditioning Psychology 3306.
Copyright © 2011 Pearson Education, Inc. All rights reserved. Developing Behavioral Persistence Through the Use of Intermittent Reinforcement Chapter 6.
Instrumental Learning A general class of behaviors inferring that learning has taken place.
Operant Conditioning. Shaping shaping = successive approximations toward a goal a process whereby reinforcements are given for behavior directed toward.
Learning Operant Conditioning.  Operant Behavior  operates (acts) on environment  produces consequences  Respondent Behavior  occurs as an automatic.
Myers EXPLORING PSYCHOLOGY (6th Edition in Modules) Module 19 Operant Conditioning James A. McCubbin, PhD Clemson University Worth Publishers.
Chapter 8 Operant Conditioning.  Operant Conditioning  type of learning in which behavior is strengthened if followed by reinforcement or diminished.
Introduction to Operant Conditioning. Operant & Classical Conditioning 1. Classical conditioning forms associations between stimuli (CS and US). Operant.
More Instrumental (Operant) Conditioning. B.F. Skinner Coined the term ‘Operant conditioning’ Coined the term ‘Operant conditioning’ The animal operates.
PSY402 Theories of Learning Chapter 4 (Cont.) Schedules of Reinforcement.
Schedules of Reinforcement Lecture 14. Schedules of RFT n Frequency of RFT after response is important n Continuous RFT l RFT after each response l Fast.
PSY 402 Theories of Learning Chapter 7 – Behavior & Its Consequences Instrumental & Operant Learning.
PSY 402 Theories of Learning Chapter 7 – Behavior & Its Consequences Instrumental & Operant Learning.
Lectures 15 & 16: Instrumental Conditioning (Schedules of Reinforcement) Learning, Psychology 5310 Spring, 2015 Professor Delamater.
OPERANT CONDITIONING DEF: a form of learning in which responses come to be controlled by their consequences.
Learning the Consequences of Behavior
Chapter 7 Operant Conditioning:
Week 5: Increasing Behavior
Ratio Schedules Focus on the number of responses required before reinforcement is given.
Psychology of Learning EXP4404 Chapter 6: Schedules of Reinforcement Dr. Steve.
Chapter 9 Adjusting to Schedules of Partial Reinforcement.
Operant Conditioning: Schedules and Theories Of Reinforcement.
Learning Prof. Tom Alloway. Definition of Learning l Change in behavior l Due to experience relevant to what is being learned l Relatively durable n Conditioning.
Chapter 6 Operant Conditioning Schedules. Schedule of Reinforcement Appetitive outcome --> reinforcement –As a “shorthand” we call the appetitive outcome.
Ninth Edition 5 Burrhus Frederic Skinner.
Operant Conditioning: Schedules and Theories of Reinforcement
B.F. SKINNER - "Skinner box": -many responses -little time and effort -easily recorded -RESPONSE RATE is the Dependent Variable.
Organizational Behavior Types of Intermittently Reinforcing Behavior.
Chapter 3 Learning (II) Operant (Instrumental) Conditioning.
Classical Conditioning
Chapter 13: Schedules of Reinforcement
Increasing & Decreasing Behaviors 1. Increasing Behaviors 2.
Chapter 6 Developing Behavioral Persistence Through the Use of Intermittent Reinforcement.
Operant Conditioning E.L. Thorndike and B.F. Skinner.
PSY402 Theories of Learning Chapter 6 – Appetitive Conditioning.
Principles of Behavior Sixth Edition Richard W. Malott Western Michigan University Power Point by Nikki Hoffmeister.
Operant Conditioning. Operant Conditioning – A form of learning in which voluntary responses come to be controlled by their consequences. What does this.
Instrumental/Operant Conditioning. Thorndike’s Puzzle Box.
Operant conditioning (Skinner – 1938, 1956)
Schedules of Reinforcement Thomas G. Bowers, Ph.D.
Schedules of Reinforcement or Punishment: Ratio Schedules
Schedules of reinforcement
Maintaining Behavior Change Dr. Alan H. Teich Chapter 10.
Principles of Behavior Sixth Edition Richard W. Malott Western Michigan University Power Point by Nikki Hoffmeister.
Schedules of Reinforcement
Operant Conditioning I. Volunteer? Priscilla the Fastidious Pig
Operant Conditioning Chapter 6.
How is behavior “shaped” through operant conditioning? Operant shaping: demonstration, analysis and terminology Chaining operant behavior (again) The cumulative.
1 Quiz Question: In what way are fixed-ratio (FR) and variable-ratio (VR) reinforcement schedules: (a) similar? (b) different?
Learning Principles & Applications 7-9% of AP Exam.
Schedules of Reinforcement
Seminar 4 Applied Behavior Analysis I PS 360 Israel A. Sarasti, Ph.D.
Operant Conditioning First Hour – How is behaviour shaped by operant conditioning? Paradigms of Operant Conditioning Edward Thorndike The Law of Effect.
Schedules and more Schedules
Operant or Instrumental Conditioning
Factors Affecting Performance on Reinforcement Schedules
Schedules of Reinforcement
Maintaining Behavior Change Chapter 10
UNIT 4 BRAIN, BEHAVIOUR & EXPERIENCE
Operant Conditioning, Continued
Schedules of Reinforcement
Presentation transcript:

Schedules of Reinforcement CH 17,18,19

Divers of Nassau Diving for coins Success does not follow every attempt Success means reinforcement

Intermittent Reinforcement A reinforcer follows the response only once in a while.

Continuous Reinforcement (CRF) A reinforcer follows every response. Best used for shaping or maintaining difficult behavior.

Shaping with Reinforcement Before: Andrew has no gum Behavior Initial: NA Intermed: Moves lips only Terminal: Says words unclearly Behavior Initial: Andrew moves lips Intermed: Andrew makes croaking sound Terminal: Says words clearly After: Andrew receives no gum After: Andrew receives gum Reinforcement Extinction

Schedule of Reinforcement The way reinforcement occurs because of the number of responses, time between responses, and stimulus conditions.

Schedules of Reinforcement Ratio –Fixed Ratio (FR) –Variable Ratio (VR) Interval –Fixed Interval (FI) –Variable Interval (VI)

Fixed Ratio (FR) Schedule of Reinforcement A reinforcer follows a fixed number of responses.

FR Responding After a response is reinforced, no responding occurs for a period of time, then responding occurs at a high, steady rate until the next reinforcer is delivered.

Pauses Reinforcers Reinforcement always immediately follows the 5 th response!!!! This is an FR5 schedule of reinforcement.

Postreinforcement Pause (PRP) The name for the pause after the consumption of the reinforcer and before the next ratio of responses begins.

FR & PRP Length of the pause is proportional to the size of the ratio. Small ratio – pausing is short Large ratio – pausing is long

General Rule for establishing intermittently reinforced behavior: First use CRF and gradually increase the intermittency of reinforcement as responding stabilizes at a high rate.

FR120 What are the response requirements? When will reinforcement be delivered? What will the pattern of responding look like?

Cumulative Graph for FR Reinforcers Post reinforcement pauses

Variable Ratio (VR) Schedule A reinforcer follows after a variable number of responses

VR Responding VR schedules produce a high, constant rate of responding, with almost no postreinforcement pausing.

VR 50

Schedule notation Number designates the average number of responses required for reinforcement

How do typical schedules of reinforcement differ from gambling? See pages

Skinner Box vs. 1 Armed Bandit Typical VRGambling Schedule Many interspersed learned reinforcers NoYes Amount of reinforcer varies from ratio to ratio NoYes Small ratio NoYes Emotional reinforcers NoYes

Time Dependent Schedules Fixed Interval (FI) Variable Interval (VI)

FI Schedule of Reinforcement A reinforcer is contingent on the first response after a fixed interval of time since the last opportunity for reinforcement.

Example: Fixed Interval (FI) Schedule of Reinforcement FI 200-s 200 seconds 400 seconds 600 seconds

Fixed-Interval Scallop A FI schedule often produces a scallop – a gradual increase in the rate of responding with responding occurring at a high rate just before reinforcement is available. No responding occurs for some time after reinforcement (Post Reinforcement Pause.

Example: Fixed Interval (FI) Schedule of Reinforcement FI 200-s 200 seconds 400 seconds 600 seconds Scallop PRP

Lever Pressing of Mice after Training

Term paper vs. FI Term paper does not have a deadline.

FI vs. Term Paper FITerm Paper Does early responding effect anything? NoYes Do you get more if you work harder? NoYes Is the relevant response class clear? YesNo Are there calendars and clocks? NoYes Is there a deadline? NoYes Is the reinforcer too delayed? NoYes

Fixed Time & Superstitious Behavior Fixed time schedules of reinforcement –A reinforcer is delivered after the passage of a fixed period of time, independently of the response. Superstitious behavior –Behaving as if the response causes some specific outcome when it really does not.

Superstitious behavior of the pigeon: Experiment by Skinner Fixed Time Schedule

Variable Interval (VI) Schedule of Reinforcement A reinforcer is contingent on the first response after a variable interval of time since the last opportunity for reinforcement.

VI Responding VI schedules produce a moderate rate of responding with almost no postreinforcement pausing.

Comparing/Contrasting Ratio and Interval Schedules Pg. 305: table Pg. 305: cumulative records

Continuous vs. Intermittent CRF – every response is reinforced Intermittent reinforcement – only some responses produce a reinforcer.

Intermittent Reinforcement Makes the response more resistant to extinction than does continuous reinforcement.

Resistance to Extinction The number of responses or the amount of time before a response extinguishes

Resistant to Extinction and Intermittent Reinforcement Intermittent reinforcement makes the response more resistant to extinction than does continuous reinforcement. Why? Hard for the rat to tell the difference between only an occasional reinforcement and no reinforcement CRF followed by EXT: rats quickly stop responding: easy discrimination between reinforcement & extinction contingencies

4 Types of Concurrent Contingencies 1.Two compatible responses 2.Compatible contingencies for a single response 3.Incompatible contingencies for a single response 4.Two incompatible responses

Concurrent Contingencies Contingencies that are available at the same time. More than one contingency of reinforcement or punishment is available at the same time.

Concurrent Contingencies Two levers in chamber –VI 300 s on Left lever –VI 30 s on Right lever Will animal allocate more responses to the left or the right lever? Matching Law (pg. 330) % of Left lever presses = % Left-lever reinforcers

Matching Law Relative Frequency of responding on two concurrent schedules of reinforcement equals the relative value of reinforcement on those two schedules