Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
112 Cards in this Set
- Front
- Back
______________ and Operant Learning can be used interchangeably
|
Instrumental and Operant learning will be used interchangeably.
|
|
- behavior can be repeated any # of times
- dependent variable is usually the # of times behavior occurs per unit of time (e.g. skinner’s box) |
Free Operant Procedures
|
|
Discrete Trial Procedure
|
-behavior of participant ends trial (e.g. escaping puzzle or rats in a maze)
|
|
Primary Reinforcer
|
- is one that is thought to be naturally or innately reinforcing (praise, food)
|
|
Secondary Reinforcer
|
is one that is dependent on its association with other reinforcers
|
|
Pre-Mac Principle
|
a highly preferred activity can be used as a reinforcer for a less preferred activity
|
|
Reinforcement
|
consequence that increases that a behavior will be repeated
|
|
consequence that decreases the probability that a behavior will be repeated
|
Punishment
|
|
Positive Contingency (+)
|
introduce stimulus after the behavior is performed
|
|
Negative Contingency (-)
|
remove stimulus after the behavior is performed
|
|
Skinner believed most effective way to bring about behavior was ___________ _____________.
|
Skinner believed most effective way to bring about behavior was positive reinforcement.
|
|
Shaping
|
is a technique that involves reinforcing successive approximations of a desired behavior
|
|
Chaining
|
training a sequence of behaviors in which the reinforcement for one behavior is the opportunity to perform the next behavior
|
|
What is a drive according to Hull?
|
Hull says a drive is a motivation (such as hunger) cause by a period of deprivation (such as food)
|
|
attributes a reinforcers effectiveness to the reduction of a drive.
|
Hull's Drive Reduction Theory of Reinforcement
|
|
What is a reinforcer? (according to Hull)
|
is a stimulus that reduces one or more drives
|
|
Most scientists find hulls Drive reduction theory unsatisfactory? or satisfactory? and why?
|
doesn't account for reinforcers that....
1) neither reduce drives 2) nor acquire their reinforcing properties from their association with primary reinforcers |
|
What three drives are covered by the alternative drive theory?
|
1) Drive of Curiosity (Novelty Seeking)
2) Activity Drive 3) Contact Comfort Drive |
|
What is preMack's relative value theory?
|
Different kinds of behavior have different values that are relative to one another at any given moment, and those relative values determine the reinforcing properties of behavior
|
|
if a person shows a stronger inclination to PARTY than to WORK, partying can reinforce working.
What principle? |
Premack principle
|
|
_____________ ____________ ______________ explain what neither the absolute nor relative theories explain
|
Opponent Process Theories
|
|
focuses more on engaging in behavior to get a positive outcome that you’ve been deprived of. Emphasis is on whether or not you’ve been pulled down from baseline. This is a different take on expanding relative value theory
|
Response Deprivation Theory
|
|
Explains risky or adrenaline seeking behaviors, specifically seeking something out.
|
Opponent Process Theory
|
|
Which theories explain avoidance?
|
The One and Two Process Theories
|
|
Response Deprivation Theory
|
-A behavior becomes reinforcing when the individual is prevented from engaging in it at its normal frequency
-If the rate of the behavior falls below the baseline level the opportunity to engage in that behavior will be reinforcing |
|
often referred to as Stimulus Response (SR) learning
|
Thorndike's Law of Effect
|
|
What did Skinner believe was the most effective way to bring about a behavior?
|
positive reinforcement
|
|
What did Skinner believe was the most effective way to bring about a behavior?
|
positive reinforcement
|
|
Hull's Drive Reduction Theory works well explaining primary reinforcers such as food or water because those reinforcers ___________________________________.
|
alter a physiological state.
|
|
Most scientists find Hull's Drive Reduction theory unsatisfactory because...
|
there are too many reinforcers that neither reduce drives nor acquire their reinforcing properties from their association with primary reinforcers
|
|
Which theory views behavior as a reinforcer, does not depend on the distinction between primary and secondary reinforcers and has no assumed physiological drives?
|
The Relative Value Theory and Pre-Mack Principle
|
|
Premack Principle
|
high probability behavior reinforcers low probability behavior
|
|
What are the two theories of avoidance?
|
1) two process theory
2) one process theory |
|
negative reinforcement
|
if something increases the probability of behavior and what you are doing is avoiding or removing something
|
|
often called omission training. It is when you remove a pleasant stimulus to decrease the probability of behavior
ex: get grounded, get phone taken away |
negative punishment
|
|
Two Process Theory
|
- avoidance learning consists of classical and operant learning
- an aversive CS will reduce the rate of ongoing behavior (conditioned suppression) -avoidance behavior will begin to extinguish once learned. |
|
Avoidance learning only involves Operant Learning in what theory?
|
One Process Theory
|
|
reducing aversive stimulation is a reinforcer for escape and avoidance behaviors
|
One Process Theory
|
|
The Behavioral Regulation Theory
|
animals have a preferred level of every behavior they engage in. When a behavior is blocked or prevented so that it is deprived below its preferred baseline level, access to it becomes reinforcing. Behavior regulation theory replaces the Premack principle as a way of identifying potential reinforcers.
|
|
What is a bliss point?
|
amount of time an individual will spend freely doing an activity
- the higher the bliss point, the more a person values that activity. -the blisspoints for different activities thus provide an objective scale of liking or value. |
|
Schedule of Reinforcement
|
a program or rule that determines how and when the occurrence of a response will be followed by a reinforcer
|
|
Continuous Reinforcement (CRF)
|
A certain behavior is reinforced every single time it occurs
|
|
-simplest of the simple schedules
-leads to very rapid increases in rate of behavior -very useful when shaping new behavior or behavior chain -rare/impractical in natural environments |
Continuous Reinforcement (CRF)
|
|
Partial or Intermittent Reinforcement
|
responding is only reinforced some of the time
|
|
Ratio Schedules
|
reinforcement depends on the # of responses the organism has performed
|
|
Interval Schedules
|
responses reinforced only if responses occur after a certain amount of time has passed.
|
|
What is a special case of ratio schedule?
|
CRF (Continuous Reinforcement) - only one response is required for reinforement
|
|
A set and consistent # of responses are required before reinforcement
|
Fixed Ratio Schedule
|
|
After you dial 10 buttons for a call to go through you are reinforced by the call going through
|
Fixed Ratio Schedule (FR,10)
|
|
People who work in factories are paid for the # of pieces they make.
Which schedule of reinforcement is this an example of? |
Fixed Ratio Schedule (Fr, ___)
|
|
Ratio Run
|
high and steady rate of responding that completes each ratio requirement
|
|
Run Rate
|
time to complete the ratio run
(the rate at which a behavior occurs once it has resumed following reinforcement) |
|
Post Reinforcement Pause
|
zero rate of responding that occurs just after reinforcement (length is greater for higher ratios)
-occurs in a fixed ratio schedule |
|
The opposite of continuous reinforcement is __________.
|
The opposite of continuous reinforcement is EXTINCTION. (Can be thought of as a schedule of nonreinforcement)
|
|
Fixed Ratio Schedule
The more work required for each reinforcement the (shorter/longer) the post reinforcement pause? |
The more work required for each reinforcement the LONGER the post reinforcement pause.
|
|
The number of responses required for reinforcement varies around an average.
|
Variable Ratio Schedule.
|
|
Slot Machines - you don't know how many times you are going to pull the slot machine down until you get the reward. This is an example of what type of schedule?
|
Variable Ratio Schedule
|
|
Fixed Interval Schedule
|
the behavior under study is reinforced the first time it occurs after a constant interval
|
|
Stretching the Ratio
|
increasing required number of responses
ex: The experimenter might start with a CRF Schedule, and then when the animal is working at a steady rate, increase the ratio to FR3; when this schedule has been in force for a while the experimenter may go to FR 5, then FR 8, then FR 20, FR30, and so on. |
|
Ratio Strain
|
-moving from a low ratio requirement (dense schedule) to a high ratio requirement (lean schedule)
- a disruption in responding due to an overly demanding response requirement -if stretching the ratio is done too fast too far, the tendency to perform will break down |
|
procedure in which you withhold the reinforcers that maintain the behavior
|
extinction
|
|
Spontaneous Recovery
|
resurgence of previously reinforced behaviors when the extinction procedure is interrupted
|
|
_________ schedules produce fewer and shorter postreinforcement pauses than _____ schedules.
|
VR schedules produce fewer and shorter postreinforcement pauses than FR schedules
|
|
As fast as the cheetah is, it does not bring down a victim every time it gives chase, nor can it depend on being successful on the second, third, or fourth try. There is no predicting which particular effort will be successful. It may be successful on two succeeding attempts, and then it may fail on the next ten tries. All that can be said is that, on average, one in every so many attempts will be reinforced.This is an example of what type of reinforcement schedule?
|
Variable Ratio
|
|
A salesperson working on commission, and gambling are both examples of what types of schedules?
|
Variable Ratio Schedules
|
|
Fixed Interval Scallop
|
-the fixed interval schedule produces a scallop shaped curve due to no responding after reinforcement and high rates of responding as reinforcement time approaches.
|
|
What type of behavior is being reinforced in fixed interval schedules?
|
instrumental behavior
|
|
Variable Interval Schedule
|
responses are reinforced if they occur after a variable amount of time since the delivery of the previous reinforcer
|
|
Waiting in line at the bank or theater. What type of Reinforcement Schedule is this?
|
Variable Interval Schedule
|
|
________ schedules produce higher response rates than comparable __________ schedules
|
ratio schedules produce higher response rates than comparable interval schedules (see graph)
|
|
Fixed Duration Schedule (FD)
|
reinforcement is contingent on the continuous performance of a behavior for some period of time
|
|
Variable Duration Schedule (VD)
|
the required period of performance varies around some average
|
|
Differential Reinforcement of Low Rate (DRL)
|
-response is reinforced only if it occurs after a certain amount of time has elapsed following the previous response
- interval begins again each time the behavior is performed - In a DRL 5" schedule - if the pigeon pecks before the 5 s interval is up this delays the reinforcement (Used to encourage low rates of responding. It is like an interval schedule, except that premature responses reset the time required between behavior) |
|
Differential Reinforcement of High Rates (DRH)
|
require that a behavior be performed a minimum number of times in a given period
ex: A pigeon might be required to peck a disk 5 times in a ten-second period. If it pecks fewer than five times during that period it receives nothing. |
|
_____________ schedules produce a high, constant rate of responding, with almost no postreinforcement pausing
|
VARIABLE RATIO schedules produce a high, constant rate of responding, with almost no postreinforcement pausing
|
|
– a gradual increase in the rate of responding with responding occurring at a high rate just before reinforcement is available. No responding occurs for some time after reinforcement (Post Reinforcement Pause)
|
Fixed Interval Scallop
|
|
Similar to interval schedules except that no behavior is required for reinforcement
|
Time Schedules ...Fixed and Variable
|
|
Cathy thinks she knows all the answers and is constantly raising her hand in class. The in- structor wants her to contribute but does not want her to monopolize the class discussion. She sets a limit for Cathy’s contributions at five. If Cathy limits her contributions to five during the period, she gets to be the teacher’s assistant for the next period. If she goes above five, another student is picked to be the assistant, and the teacher ignores her hand raises.
What type of reinforcement schedule is this? |
DRL
|
|
Mrs. Sadie wants Tim to contribute more in class during social studies. During this time, he seldom says a word. The teacher has indicated that Tim can have 10 minutes on the computer to play games if he will contribute at least three times during the class discussion. Once he consistently contributes at least three times, she increases the number of discussion contri- butions one per week until he is contributing at least five times per discussion period.
What type of reinforcement schedule is this? |
DRH
|
|
What is the key difference between PreMack's relative value theory and the Response deprivation theory?
|
The response deprivation theory says that the relative value of one reinforcer to another is not vital; what is vital is the extent to which each behavior occurs below its baseline rate.
|
|
Thorndike's Law of effect
|
The Strength of a behavior depends on its consequences.
If a behavior is followed by a satisfying event it is strengthened. If a behavior is followed by an annoying event the S-R association is weakened) |
|
Experiences whereby behavior is strengthened or weakened by its consequences is ________ because the behavior operates on the environment
|
OPERANT LEARNING
|
|
What is the difference between the optimum arousal theory and the Drive Reduction theory?
|
The optimal arousal theory holds that reinforcement consists not necessarily of a decrease in a drive but rather a change in the drive toward some optimal level.
|
|
A rat receives food for pressing a lever. It receives food for pressing the lever only if at least 10s have elapsed since the last lever press. What time of reinforcement schedule is this?
|
Differential Reinforcement of Low Rate (DRL)
|
|
Fixed Time Schedule (FT)
|
a reinforcer is delivered after a given period of time without regard to behavior
|
|
Variable Time Schedule (VT)
|
Reinforcement is delivered periodically at irregular intervals regardless of what behavior occurs.
|
|
Parents may praise their children each time they see them studying; gradually, however, they may reinforced the behavior less often. This is an example of what?
|
Stretching the Ratio
|
|
extinction
|
procedure in which you withhold the reinforcers that maintain behavior.
|
|
Partial Reinforcement Effect (PRE)
|
one peculiar schedule effect is the tendency of behavior that has been maintained on an intermittent schedule to be more resistant to extinction that behavior that has been on continuous reinforcement
|
|
Discrimination Hypothesis
|
Extinction takes longer after intermittent reinforcement because it is harder to distinguish (or discriminate) between extinction and an intermittent schedule than between extinction and continuous reinforcement.
|
|
Frustration Hypothesis
|
(the emotional state of frustration becomes a que for lever pressing)
Now, when the behavior is placed on extinction the animal becomes frustrated, but the frustration is a signal for lever pressing, so responding continues. |
|
Sequential Hypothesis
|
during intermittent reinforcement training, the memory of nonreward becomes a cue for performing the instrumental response, due to sequence of reinforced and nonreinforced responses.
|
|
Response Unit Hypothesis
|
When responses are defined in terms of the units required for reinforcement, the total number of responses during extinction declines as the reinforcement schedule gets thinner. When we define responses in terms of units required for reinforcement, we find that "the apparent advantage of so called intermittent reinforcement disappears". Behavior on intermittent reinforcement only seems to be more resistant to extinction because we have failed to take into account the response units required for reinforcement.
|
|
Which hypothesis is calling PRE an illusion?
|
The Response unit hypothesis
|
|
Multiple Schedule
|
two or more simple schedules that alternate and each is associated with a particular schedule, a signal indicates which schedule your on
|
|
Mixed Schedule
|
the same as multiple schedule except that there are no stimuli(or clear indication) associated with the change in reinforcement contigencies.
|
|
Chain Schedule
|
reinforcement is delivered only on completion of the last in a series of schedules. (There is a distinctive event that signals the end of one schedule and the beginning of the next)
|
|
Tandem Schedule
|
identical to chain schedule except that there is no distinctive event that signals the end of one schedule and the beginning of the next.
|
|
Concurrent Schedules
|
two or more schedules are available at once.
|
|
Matching Law
|
the relative rate of responding on a particular response alternative equals the relative rate of reinforcement for that response alternative
|
|
Cooperative Schedules
|
reinforcement is contingent on behavior of two or more subjects.
In this type of schedule, the reinforcement that one subject gets is partly dependent on the behavior of the other subject. |
|
What are the 5 problems associated with punishment?
|
Escape, Aggression, Apathy, Abuse, and Imitation of the punisher.
|
|
Response Prevention
|
Prevent the behavior from occurring by altering the environment in some way
|
|
Differential Reinforcement of Incompatible Behavior (DRI)
|
you reinforce a behavior that is incompatible with the unwanted behavior.
|
|
Punishment - contingency
|
the degree of contingency between a behavior and a punishing event. The extent to which the behavior results in a consequence. The closer to 100% the stronger the effects of punishment
|
|
Punishment - Contiguity
|
how close in time the punishment follows the behavior (immediate = most effective)
|
|
Punisher Intensity
|
The strength of the punisher is an important factor in its effects on behavior.
-very mild punishers typically have no effect. -Stronger punisher leads to a greater reduction in the punished behavior. |
|
Introductory Level of Punishment
|
One must begin with a punisher that is intense enough to suppress the behavior dramatically
*The problem in beginning with a weak punisher and gradually increasing its intensity is that the punished behavior will tend to persist during these increases. So, in the end a far greater level of punisher may be required to suppress the behavior. |
|
Alternative Sources of Reinforcement
|
if there are alternative ways of obtaining reinforcement the punished behavior is likely to decrease; if not the punished behavior is likely to persist
|
|
Responses that are followed by reinforcement become more probable. What does this refer to?
|
Thorndike's Law of effect
|
|
the terms "positive" and "negative" in punishment refer to ____________
|
whether something is presented or removed
|
|
Stimuli that governs behavior that occurs before the response is _________________________, whereas stimuli that governs behavior that occurs after the response is ____________________.
|
Stimuli that governs behavior that occurs before the response is classical conditioning whereas stimuli that governs behavior that occurs after the response is operant learning.
|