• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/96

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

96 Cards in this Set

  • Front
  • Back
Continuous Reinforcement
Behavior that is reinforced after every time it occurs.
Actual continuous reinforcement occurs infrequently.
Speed of Extinction for Continuous Reinforcement.
Quick
Partial reinforcement
behavior that is reinforced some but not all of the time. (not after every single occurence)
Speed of Extinction for Partial reinforcement
takes longer than continuous reinforcement
Types of partial reinforcement
fixed-interval, variable-interval, fixed-ratio, variable-ratio
Fixed-interval
a fixed amount of time goes by between reinforcements
Example of fixed-interval
tests recorded on schedule, so fixed.
length of post-reinforcement pauses for fixed interval
long
variable-interval
a randomly changing amount of time goes by between reinforcements
examples of variable-interval
an animal waiting for its prey, fishing.
Snake waits for frog, misses, waits again, eventually he'll eat, doesn't know when.
Pop quiz - don't know when, study more.
Difference between fixed-interval and fixed-time
fixed-interval involves reinforcing behavior, fixed time does not require certain behavior before reinforcement, only depends on certain time passing.
fixed-ratio
reinforcement occurs after a fixed number of correct responses
example of fixed-ratio
factory worker gets paid $10.00 for every 20 widgets.
length of post-reinforcement pauses for fixed-ratio
present; once you finish the job, you get reinforced, then take small break
variable-ratio
reinforcement occurs after a changing number of correct responses. VR5 = # of responses on average that occur before reinforcement.
example of variable-ratio
slot machines
might be VR 2000, don't know when you'll be reinforced, so keep going.
length of reinforcement pauses for variable-ratio
very few, very brief
Simple schedules
fixed duration, variable duration, differential reinforcement high, differential reinforcement low, fixed time, variable time
fixed duration
continuous performance for some period of time
example of fixed duration
exercise for 30 minutes straight
variable duration
required period of performance varies around an average
example of variable duration
performance has to occur after a period of time, rewared for practicing instrument while mom making dinner.
differential reinforcement high
a behavior must be performed a minimum number of times in a given period
example of differential reinforcement high
read at least 5 times this week for homework
differential reinforcement low
a behavior must occur no more than maximum amount of times in a given period
example of differential reinforcement low
must not bite nails more than 5 times today.
fixed time
reinforcer is given after a given period of time has passed, regardless of behavior
example of fixed time
children get recess regardless of work done in class
variable time
reinforcer is given at irregular intervals, regardless of behavior
example of variable time
receive candy from baby sitter sometimes
Simple schedules that can lead to superstition
fixed time and variable time; because the reinforcer may be believed to be for the current behavior even though it is not for any behavior
Partial reinforcement effect
the tendency for a behavior that has been maintained on an intermittent schedule to be more resistant to extinction than a behavior that has been on continuous reinforcement.
Explanations for PRE (partial reinforcement effect)
Discrimination hypothesis, frustration hypothesis, sequential hypothesis, or response unit hypothesis
Discrimination hypothesis
extinction takes longer after intermittent reinforcement because it is harder to discriminate between extinction and an intermittent schedule than between extinction and continuous reinforcement.
FR30 harder to discriminate than FR1.
Frustration Hypothesis
nonreinforcement of previously reinforced behavior is frustrating (aversive emotional state), so anything that reduces frustration will be reinforcing. No frustration in continuous reinforcement because no nonreinforcement. The thinner the reinforcement schedule, the higher the level of frustration.
Sequential Hypothesis
differences in the sequence of cues during training. some nonreinforcement is to be expected, the thinner the reinforcement schedule, the more resistant the rat will be to extinction, since it is used to a long period of nonreinforcement.
The frustration and sequential hypotheses are variations of the _____ hypothesis
discrimination
response unit hypothesis
to understand the PRE, we must think differently about the behavior on intermittent reinforcement. The behavior is defined by the number of times it must occur to produce reinforcement. The total number of responses declines as the reinforcement schedule gets thinner. THE PRE IS AN ILLUSION, BEHAVIOR ON INTERMITTENT REINFORCEMENT ONLY SEEMS TO BE MORE RESISTANT TO EXTINCTION BECAUSE WE HAVE FAILED TO TAKE INTO ACCOUNT THE RESPONSE UNITS REQUIRED FOR REINFORCEMENT.
complex schedules
multiple schedule, mixed schedule, chain schedule, tandem schedule, cooperative schedule, concurrent schedule,
multiple schedule
a behavior is under the influence of two or more simple schedules, each associated with a particular stimulus
mixed schedule
a behavior is under the influence of two or more simple schedules, without stimulus
chain schedule
reinforcement is delivered only on completion of the last in a series of schedules.
tandem schedule
reinforcement is delivered only on completion of the last in a series of schedules, with no distinction between the end of one schedule and the beginning of another schedule.
cooperative schedule
schedules that make reinforcement dependent on the behavior of two or more individuals.
concurrent schedules
two or more schedules are available at once, involves choice.
matching law
involved in choice, the distribution of behaviors matches the availability of reinforcement.
applications of schedules
compulsive gambling, experimental economics, malingering
compulsive gambling as application of schedules
variations in the schedule: effects of early wins,
near misses make a substantial difference
experimental economics as application of schedules
when the price of a luxury item rises, the consumption of that item declines. when the price of an essential item rises, little change in consumption.
malingering as application of schedules
pretending to be in pain in order to avoid unpleasant duties.
operant behavior association with chronic pain may be maintained by reinforcement after the pain has ceased. People will malinger longer if others are willing to press the lever.
schedules of reinforcement
the rule describing the delivery of reinforcement
schedule effects
a particular kind of reinforcement schedule tends to produce a particular pattern and rate of performance
post-reinforcement pauses
the pause that follows reinforcement
run rate
the rate at which the organism performs once it has resument work after reinforcement.
stretching the ratio
successive approximations of the desired behavior are reinforced. FR5, then FR8, then FR12, etc.
forgetting
deterioration in learned behavior following a period without practice
measuring forgetting
free recall, prompted/cued recall, relearning method, savings method, delayed matching to sample
retention interval
a period during which the learned behavior is not performed, before measurement occurs
free recall
the organism is given the opportunity to perform a previously learned behavior following a retention interval
prompted/cued recall
presented hints, to increase the likelihood that behavior will be produced
relearning method
assumes that the less training required to reach the previous level of performance, the less forgetting has occurred.
savings method
a savings of information compared to the original training program
recognition
the participant has only to identify the material previously learned.
delayed matching to sample
the animal is prevented from performing following presentation of the sample.
extinction method
immediate extinction training vs. training following removal from the chamber. if extinction proceeds more rapidly after the retention interval, fogetting has occurred
gradient degradation
forgetting may be measured as a flattening of a generalization gradient. a generalization gradient indicates forgetting.
variables in forgetting
passage of time: the longer the interval between training, the greated the deterioration;
time is not an event, so can't cause other events.
DEGREES OF LEARNING, overlearning, PRIOR LEARNING, proactive interference, paired associate learning, SUBSEQUENT LEARNING, retroactive interference, CONTEXT, cue-dependent forgetting, reminiscence, dependent learning
Overlearning
practicing even beyond mastery
fluency
number of correct responses per minute while learning
proactive interference
old learning can interfere with recall
paired associate learning
used to study proactive interference; the person learns a list of word pairs, so that when given the first word, can recall the second.
retroactive interference
when what we learn interferes with our ability to recall earlier learning.
cue-dependent forgetting
learning occured with presence of a particular stimuli and that stimuli is now absent.
reminiscence
when performance improves with the passage of time
state-dependent learning
behavior that is learning during a particular physiological state is lost when that state passes.
foraging
the forgetting curve may be less steep when what is learned is important to survival.
eyewitness testimony
dependent on way questions are phrased
ways of overlearning
mnemonics, method of loci, peg word system,
mnemonic
any device for aiding recall, ex: rhyme, acronym
mnemonic systems
method of loci, peg word system
method of loci
apply grocery list to an imaginary walk through a place you frequently visit.
peg word system
one is a bun, two is a shoe
use concret objects as pegs.
context clues
study in the classroom
prompts
note on calendar
synesthesia
a synthesis of different senses; sound produces light, color, taste, odor, touch
limits on learning
physical structure: incapable of speaking; learned behavior is not inheritable, dies with the learner
Lamarck
theorized that physical characteristics were acquired adaptations that were passed on from generation to generation.
heredity
plays important role in human learning abilities within and between species.
environment
enriched environments may create intelligence
neurological damage
prenatal exposure to alcohol results in limited learning abilities;
neurotoxin damage nerve tissues;
head injury, blow to head or shaking of child;
disease and malnutrition: prevent normal neurological development
critical periods
stages that are optimum for learning. certain opportunities may occur once in a lifetime.
imprinting
Lorenz (1952) the youngster will become attached to any moving object that happens to pass by, whether another animal of the same species, a mechanical object, or a human being
Breland (1961)
innate tendencies interfere with learning
raccoon not letting go of coin
Garcia and Koelling (1966)
taste aversion
continuum of preparedness
an organism comes to a learning situation genetically prepared to learn(quick), unprepared(steady but slow), or contraprepared(slow and irregular)
Seligman (1970)
continuum of preparedness