• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/112

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

112 Cards in this Set

  • Front
  • Back
______________ and Operant Learning can be used interchangeably
Instrumental and Operant learning will be used interchangeably.
- behavior can be repeated any # of times
- dependent variable is usually the # of times behavior occurs per unit of time (e.g. skinner’s box)
Free Operant Procedures
Discrete Trial Procedure
-behavior of participant ends trial (e.g. escaping puzzle or rats in a maze)
Primary Reinforcer
- is one that is thought to be naturally or innately reinforcing (praise, food)
Secondary Reinforcer
is one that is dependent on its association with other reinforcers
Pre-Mac Principle
a highly preferred activity can be used as a reinforcer for a less preferred activity
Reinforcement
consequence that increases that a behavior will be repeated
consequence that decreases the probability that a behavior will be repeated
Punishment
Positive Contingency (+)
introduce stimulus after the behavior is performed
Negative Contingency (-)
remove stimulus after the behavior is performed
Skinner believed most effective way to bring about behavior was ___________ _____________.
Skinner believed most effective way to bring about behavior was positive reinforcement.
Shaping
is a technique that involves reinforcing successive approximations of a desired behavior
Chaining
training a sequence of behaviors in which the reinforcement for one behavior is the opportunity to perform the next behavior
What is a drive according to Hull?
Hull says a drive is a motivation (such as hunger) cause by a period of deprivation (such as food)
attributes a reinforcers effectiveness to the reduction of a drive.
Hull's Drive Reduction Theory of Reinforcement
What is a reinforcer? (according to Hull)
is a stimulus that reduces one or more drives
Most scientists find hulls Drive reduction theory unsatisfactory? or satisfactory? and why?
doesn't account for reinforcers that....
1) neither reduce drives
2) nor acquire their reinforcing properties from their association with primary reinforcers
What three drives are covered by the alternative drive theory?
1) Drive of Curiosity (Novelty Seeking)
2) Activity Drive
3) Contact Comfort Drive
What is preMack's relative value theory?
Different kinds of behavior have different values that are relative to one another at any given moment, and those relative values determine the reinforcing properties of behavior
if a person shows a stronger inclination to PARTY than to WORK, partying can reinforce working.

What principle?
Premack principle
_____________ ____________ ______________ explain what neither the absolute nor relative theories explain
Opponent Process Theories
focuses more on engaging in behavior to get a positive outcome that you’ve been deprived of. Emphasis is on whether or not you’ve been pulled down from baseline. This is a different take on expanding relative value theory
Response Deprivation Theory
Explains risky or adrenaline seeking behaviors, specifically seeking something out.
Opponent Process Theory
Which theories explain avoidance?
The One and Two Process Theories
Response Deprivation Theory
-A behavior becomes reinforcing when the individual is prevented from engaging in it at its normal frequency
-If the rate of the behavior falls below the baseline level the opportunity to engage in that behavior will be reinforcing
often referred to as Stimulus Response (SR) learning
Thorndike's Law of Effect
What did Skinner believe was the most effective way to bring about a behavior?
positive reinforcement
What did Skinner believe was the most effective way to bring about a behavior?
positive reinforcement
Hull's Drive Reduction Theory works well explaining primary reinforcers such as food or water because those reinforcers ___________________________________.
alter a physiological state.
Most scientists find Hull's Drive Reduction theory unsatisfactory because...
there are too many reinforcers that neither reduce drives nor acquire their reinforcing properties from their association with primary reinforcers
Which theory views behavior as a reinforcer, does not depend on the distinction between primary and secondary reinforcers and has no assumed physiological drives?
The Relative Value Theory and Pre-Mack Principle
Premack Principle
high probability behavior reinforcers low probability behavior
What are the two theories of avoidance?
1) two process theory
2) one process theory
negative reinforcement
if something increases the probability of behavior and what you are doing is avoiding or removing something
often called omission training. It is when you remove a pleasant stimulus to decrease the probability of behavior
ex: get grounded, get phone taken away
negative punishment
Two Process Theory
- avoidance learning consists of classical and operant learning
- an aversive CS will reduce the rate of ongoing behavior (conditioned suppression)
-avoidance behavior will begin to extinguish once learned.
Avoidance learning only involves Operant Learning in what theory?
One Process Theory
reducing aversive stimulation is a reinforcer for escape and avoidance behaviors
One Process Theory
The Behavioral Regulation Theory
animals have a preferred level of every behavior they engage in. When a behavior is blocked or prevented so that it is deprived below its preferred baseline level, access to it becomes reinforcing. Behavior regulation theory replaces the Premack principle as a way of identifying potential reinforcers.
What is a bliss point?
amount of time an individual will spend freely doing an activity

- the higher the bliss point, the more a person values that activity.
-the blisspoints for different activities thus provide an objective scale of liking or value.
Schedule of Reinforcement
a program or rule that determines how and when the occurrence of a response will be followed by a reinforcer
Continuous Reinforcement (CRF)
A certain behavior is reinforced every single time it occurs
-simplest of the simple schedules
-leads to very rapid increases in rate of behavior
-very useful when shaping new behavior or behavior chain
-rare/impractical in natural environments
Continuous Reinforcement (CRF)
Partial or Intermittent Reinforcement
responding is only reinforced some of the time
Ratio Schedules
reinforcement depends on the # of responses the organism has performed
Interval Schedules
responses reinforced only if responses occur after a certain amount of time has passed.
What is a special case of ratio schedule?
CRF (Continuous Reinforcement) - only one response is required for reinforement
A set and consistent # of responses are required before reinforcement
Fixed Ratio Schedule
After you dial 10 buttons for a call to go through you are reinforced by the call going through
Fixed Ratio Schedule (FR,10)
People who work in factories are paid for the # of pieces they make.

Which schedule of reinforcement is this an example of?
Fixed Ratio Schedule (Fr, ___)
Ratio Run
high and steady rate of responding that completes each ratio requirement
Run Rate
time to complete the ratio run

(the rate at which a behavior occurs once it has resumed following reinforcement)
Post Reinforcement Pause
zero rate of responding that occurs just after reinforcement (length is greater for higher ratios)

-occurs in a fixed ratio schedule
The opposite of continuous reinforcement is __________.
The opposite of continuous reinforcement is EXTINCTION. (Can be thought of as a schedule of nonreinforcement)
Fixed Ratio Schedule

The more work required for each reinforcement the (shorter/longer) the post reinforcement pause?
The more work required for each reinforcement the LONGER the post reinforcement pause.
The number of responses required for reinforcement varies around an average.
Variable Ratio Schedule.
Slot Machines - you don't know how many times you are going to pull the slot machine down until you get the reward. This is an example of what type of schedule?
Variable Ratio Schedule
Fixed Interval Schedule
the behavior under study is reinforced the first time it occurs after a constant interval
Stretching the Ratio
increasing required number of responses

ex: The experimenter might start with a CRF Schedule, and then when the animal is working at a steady rate, increase the ratio to FR3; when this schedule has been in force for a while the experimenter may go to FR 5, then FR 8, then FR 20, FR30, and so on.
Ratio Strain
-moving from a low ratio requirement (dense schedule) to a high ratio requirement (lean schedule)
- a disruption in responding due to an overly demanding response requirement
-if stretching the ratio is done too fast too far, the tendency to perform will break down
procedure in which you withhold the reinforcers that maintain the behavior
extinction
Spontaneous Recovery
resurgence of previously reinforced behaviors when the extinction procedure is interrupted
_________ schedules produce fewer and shorter postreinforcement pauses than _____ schedules.
VR schedules produce fewer and shorter postreinforcement pauses than FR schedules
As fast as the cheetah is, it does not bring down a victim every time it gives chase, nor can it depend on being successful on the second, third, or fourth try. There is no predicting which particular effort will be successful. It may be successful on two succeeding attempts, and then it may fail on the next ten tries. All that can be said is that, on average, one in every so many attempts will be reinforced.This is an example of what type of reinforcement schedule?
Variable Ratio
A salesperson working on commission, and gambling are both examples of what types of schedules?
Variable Ratio Schedules
Fixed Interval Scallop
-the fixed interval schedule produces a scallop shaped curve due to no responding after reinforcement and high rates of responding as reinforcement time approaches.
What type of behavior is being reinforced in fixed interval schedules?
instrumental behavior
Variable Interval Schedule
responses are reinforced if they occur after a variable amount of time since the delivery of the previous reinforcer
Waiting in line at the bank or theater. What type of Reinforcement Schedule is this?
Variable Interval Schedule
________ schedules produce higher response rates than comparable __________ schedules
ratio schedules produce higher response rates than comparable interval schedules (see graph)
Fixed Duration Schedule (FD)
reinforcement is contingent on the continuous performance of a behavior for some period of time
Variable Duration Schedule (VD)
the required period of performance varies around some average
Differential Reinforcement of Low Rate (DRL)
-response is reinforced only if it occurs after a certain amount of time has elapsed following the previous response

- interval begins again each time the behavior is performed

- In a DRL 5" schedule - if the pigeon pecks before the 5 s interval is up this delays the reinforcement

(Used to encourage low rates of responding. It is like an interval schedule, except that premature responses reset the time required between behavior)
Differential Reinforcement of High Rates (DRH)
require that a behavior be performed a minimum number of times in a given period

ex: A pigeon might be required to peck a disk 5 times in a ten-second period. If it pecks fewer than five times during that period it receives nothing.
_____________ schedules produce a high, constant rate of responding, with almost no postreinforcement pausing
VARIABLE RATIO schedules produce a high, constant rate of responding, with almost no postreinforcement pausing
– a gradual increase in the rate of responding with responding occurring at a high rate just before reinforcement is available. No responding occurs for some time after reinforcement (Post Reinforcement Pause)
Fixed Interval Scallop
Similar to interval schedules except that no behavior is required for reinforcement
Time Schedules ...Fixed and Variable
Cathy thinks she knows all the answers and is constantly raising her hand in class. The in- structor wants her to contribute but does not want her to monopolize the class discussion. She sets a limit for Cathy’s contributions at five. If Cathy limits her contributions to five during the period, she gets to be the teacher’s assistant for the next period. If she goes above five, another student is picked to be the assistant, and the teacher ignores her hand raises.
What type of reinforcement schedule is this?
DRL
Mrs. Sadie wants Tim to contribute more in class during social studies. During this time, he seldom says a word. The teacher has indicated that Tim can have 10 minutes on the computer to play games if he will contribute at least three times during the class discussion. Once he consistently contributes at least three times, she increases the number of discussion contri- butions one per week until he is contributing at least five times per discussion period.
What type of reinforcement schedule is this?
DRH
What is the key difference between PreMack's relative value theory and the Response deprivation theory?
The response deprivation theory says that the relative value of one reinforcer to another is not vital; what is vital is the extent to which each behavior occurs below its baseline rate.
Thorndike's Law of effect
The Strength of a behavior depends on its consequences.

If a behavior is followed by a satisfying event it is strengthened.
If a behavior is followed by an annoying event the S-R association is weakened)
Experiences whereby behavior is strengthened or weakened by its consequences is ________ because the behavior operates on the environment
OPERANT LEARNING
What is the difference between the optimum arousal theory and the Drive Reduction theory?
The optimal arousal theory holds that reinforcement consists not necessarily of a decrease in a drive but rather a change in the drive toward some optimal level.
A rat receives food for pressing a lever. It receives food for pressing the lever only if at least 10s have elapsed since the last lever press. What time of reinforcement schedule is this?
Differential Reinforcement of Low Rate (DRL)
Fixed Time Schedule (FT)
a reinforcer is delivered after a given period of time without regard to behavior
Variable Time Schedule (VT)
Reinforcement is delivered periodically at irregular intervals regardless of what behavior occurs.
Parents may praise their children each time they see them studying; gradually, however, they may reinforced the behavior less often. This is an example of what?
Stretching the Ratio
extinction
procedure in which you withhold the reinforcers that maintain behavior.
Partial Reinforcement Effect (PRE)
one peculiar schedule effect is the tendency of behavior that has been maintained on an intermittent schedule to be more resistant to extinction that behavior that has been on continuous reinforcement
Discrimination Hypothesis
Extinction takes longer after intermittent reinforcement because it is harder to distinguish (or discriminate) between extinction and an intermittent schedule than between extinction and continuous reinforcement.
Frustration Hypothesis
(the emotional state of frustration becomes a que for lever pressing)

Now, when the behavior is placed on extinction the animal becomes frustrated, but the frustration is a signal for lever pressing, so responding continues.
Sequential Hypothesis
during intermittent reinforcement training, the memory of nonreward becomes a cue for performing the instrumental response, due to sequence of reinforced and nonreinforced responses.
Response Unit Hypothesis
When responses are defined in terms of the units required for reinforcement, the total number of responses during extinction declines as the reinforcement schedule gets thinner. When we define responses in terms of units required for reinforcement, we find that "the apparent advantage of so called intermittent reinforcement disappears". Behavior on intermittent reinforcement only seems to be more resistant to extinction because we have failed to take into account the response units required for reinforcement.
Which hypothesis is calling PRE an illusion?
The Response unit hypothesis
Multiple Schedule
two or more simple schedules that alternate and each is associated with a particular schedule, a signal indicates which schedule your on
Mixed Schedule
the same as multiple schedule except that there are no stimuli(or clear indication) associated with the change in reinforcement contigencies.
Chain Schedule
reinforcement is delivered only on completion of the last in a series of schedules. (There is a distinctive event that signals the end of one schedule and the beginning of the next)
Tandem Schedule
identical to chain schedule except that there is no distinctive event that signals the end of one schedule and the beginning of the next.
Concurrent Schedules
two or more schedules are available at once.
Matching Law
the relative rate of responding on a particular response alternative equals the relative rate of reinforcement for that response alternative
Cooperative Schedules
reinforcement is contingent on behavior of two or more subjects.

In this type of schedule, the reinforcement that one subject gets is partly dependent on the behavior of the other subject.
What are the 5 problems associated with punishment?
Escape, Aggression, Apathy, Abuse, and Imitation of the punisher.
Response Prevention
Prevent the behavior from occurring by altering the environment in some way
Differential Reinforcement of Incompatible Behavior (DRI)
you reinforce a behavior that is incompatible with the unwanted behavior.
Punishment - contingency
the degree of contingency between a behavior and a punishing event. The extent to which the behavior results in a consequence. The closer to 100% the stronger the effects of punishment
Punishment - Contiguity
how close in time the punishment follows the behavior (immediate = most effective)
Punisher Intensity
The strength of the punisher is an important factor in its effects on behavior.
-very mild punishers typically have no effect.
-Stronger punisher leads to a greater reduction in the punished behavior.
Introductory Level of Punishment
One must begin with a punisher that is intense enough to suppress the behavior dramatically

*The problem in beginning with a weak punisher and gradually increasing its intensity is that the punished behavior will tend to persist during these increases. So, in the end a far greater level of punisher may be required to suppress the behavior.
Alternative Sources of Reinforcement
if there are alternative ways of obtaining reinforcement the punished behavior is likely to decrease; if not the punished behavior is likely to persist
Responses that are followed by reinforcement become more probable. What does this refer to?
Thorndike's Law of effect
the terms "positive" and "negative" in punishment refer to ____________
whether something is presented or removed
Stimuli that governs behavior that occurs before the response is _________________________, whereas stimuli that governs behavior that occurs after the response is ____________________.
Stimuli that governs behavior that occurs before the response is classical conditioning whereas stimuli that governs behavior that occurs after the response is operant learning.