Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
96 Cards in this Set
- Front
- Back
Continuous Reinforcement
|
Behavior that is reinforced after every time it occurs.
Actual continuous reinforcement occurs infrequently. |
|
Speed of Extinction for Continuous Reinforcement.
|
Quick
|
|
Partial reinforcement
|
behavior that is reinforced some but not all of the time. (not after every single occurence)
|
|
Speed of Extinction for Partial reinforcement
|
takes longer than continuous reinforcement
|
|
Types of partial reinforcement
|
fixed-interval, variable-interval, fixed-ratio, variable-ratio
|
|
Fixed-interval
|
a fixed amount of time goes by between reinforcements
|
|
Example of fixed-interval
|
tests recorded on schedule, so fixed.
|
|
length of post-reinforcement pauses for fixed interval
|
long
|
|
variable-interval
|
a randomly changing amount of time goes by between reinforcements
|
|
examples of variable-interval
|
an animal waiting for its prey, fishing.
Snake waits for frog, misses, waits again, eventually he'll eat, doesn't know when. Pop quiz - don't know when, study more. |
|
Difference between fixed-interval and fixed-time
|
fixed-interval involves reinforcing behavior, fixed time does not require certain behavior before reinforcement, only depends on certain time passing.
|
|
fixed-ratio
|
reinforcement occurs after a fixed number of correct responses
|
|
example of fixed-ratio
|
factory worker gets paid $10.00 for every 20 widgets.
|
|
length of post-reinforcement pauses for fixed-ratio
|
present; once you finish the job, you get reinforced, then take small break
|
|
variable-ratio
|
reinforcement occurs after a changing number of correct responses. VR5 = # of responses on average that occur before reinforcement.
|
|
example of variable-ratio
|
slot machines
might be VR 2000, don't know when you'll be reinforced, so keep going. |
|
length of reinforcement pauses for variable-ratio
|
very few, very brief
|
|
Simple schedules
|
fixed duration, variable duration, differential reinforcement high, differential reinforcement low, fixed time, variable time
|
|
fixed duration
|
continuous performance for some period of time
|
|
example of fixed duration
|
exercise for 30 minutes straight
|
|
variable duration
|
required period of performance varies around an average
|
|
example of variable duration
|
performance has to occur after a period of time, rewared for practicing instrument while mom making dinner.
|
|
differential reinforcement high
|
a behavior must be performed a minimum number of times in a given period
|
|
example of differential reinforcement high
|
read at least 5 times this week for homework
|
|
differential reinforcement low
|
a behavior must occur no more than maximum amount of times in a given period
|
|
example of differential reinforcement low
|
must not bite nails more than 5 times today.
|
|
fixed time
|
reinforcer is given after a given period of time has passed, regardless of behavior
|
|
example of fixed time
|
children get recess regardless of work done in class
|
|
variable time
|
reinforcer is given at irregular intervals, regardless of behavior
|
|
example of variable time
|
receive candy from baby sitter sometimes
|
|
Simple schedules that can lead to superstition
|
fixed time and variable time; because the reinforcer may be believed to be for the current behavior even though it is not for any behavior
|
|
Partial reinforcement effect
|
the tendency for a behavior that has been maintained on an intermittent schedule to be more resistant to extinction than a behavior that has been on continuous reinforcement.
|
|
Explanations for PRE (partial reinforcement effect)
|
Discrimination hypothesis, frustration hypothesis, sequential hypothesis, or response unit hypothesis
|
|
Discrimination hypothesis
|
extinction takes longer after intermittent reinforcement because it is harder to discriminate between extinction and an intermittent schedule than between extinction and continuous reinforcement.
FR30 harder to discriminate than FR1. |
|
Frustration Hypothesis
|
nonreinforcement of previously reinforced behavior is frustrating (aversive emotional state), so anything that reduces frustration will be reinforcing. No frustration in continuous reinforcement because no nonreinforcement. The thinner the reinforcement schedule, the higher the level of frustration.
|
|
Sequential Hypothesis
|
differences in the sequence of cues during training. some nonreinforcement is to be expected, the thinner the reinforcement schedule, the more resistant the rat will be to extinction, since it is used to a long period of nonreinforcement.
|
|
The frustration and sequential hypotheses are variations of the _____ hypothesis
|
discrimination
|
|
response unit hypothesis
|
to understand the PRE, we must think differently about the behavior on intermittent reinforcement. The behavior is defined by the number of times it must occur to produce reinforcement. The total number of responses declines as the reinforcement schedule gets thinner. THE PRE IS AN ILLUSION, BEHAVIOR ON INTERMITTENT REINFORCEMENT ONLY SEEMS TO BE MORE RESISTANT TO EXTINCTION BECAUSE WE HAVE FAILED TO TAKE INTO ACCOUNT THE RESPONSE UNITS REQUIRED FOR REINFORCEMENT.
|
|
complex schedules
|
multiple schedule, mixed schedule, chain schedule, tandem schedule, cooperative schedule, concurrent schedule,
|
|
multiple schedule
|
a behavior is under the influence of two or more simple schedules, each associated with a particular stimulus
|
|
mixed schedule
|
a behavior is under the influence of two or more simple schedules, without stimulus
|
|
chain schedule
|
reinforcement is delivered only on completion of the last in a series of schedules.
|
|
tandem schedule
|
reinforcement is delivered only on completion of the last in a series of schedules, with no distinction between the end of one schedule and the beginning of another schedule.
|
|
cooperative schedule
|
schedules that make reinforcement dependent on the behavior of two or more individuals.
|
|
concurrent schedules
|
two or more schedules are available at once, involves choice.
|
|
matching law
|
involved in choice, the distribution of behaviors matches the availability of reinforcement.
|
|
applications of schedules
|
compulsive gambling, experimental economics, malingering
|
|
compulsive gambling as application of schedules
|
variations in the schedule: effects of early wins,
near misses make a substantial difference |
|
experimental economics as application of schedules
|
when the price of a luxury item rises, the consumption of that item declines. when the price of an essential item rises, little change in consumption.
|
|
malingering as application of schedules
|
pretending to be in pain in order to avoid unpleasant duties.
operant behavior association with chronic pain may be maintained by reinforcement after the pain has ceased. People will malinger longer if others are willing to press the lever. |
|
schedules of reinforcement
|
the rule describing the delivery of reinforcement
|
|
schedule effects
|
a particular kind of reinforcement schedule tends to produce a particular pattern and rate of performance
|
|
post-reinforcement pauses
|
the pause that follows reinforcement
|
|
run rate
|
the rate at which the organism performs once it has resument work after reinforcement.
|
|
stretching the ratio
|
successive approximations of the desired behavior are reinforced. FR5, then FR8, then FR12, etc.
|
|
forgetting
|
deterioration in learned behavior following a period without practice
|
|
measuring forgetting
|
free recall, prompted/cued recall, relearning method, savings method, delayed matching to sample
|
|
retention interval
|
a period during which the learned behavior is not performed, before measurement occurs
|
|
free recall
|
the organism is given the opportunity to perform a previously learned behavior following a retention interval
|
|
prompted/cued recall
|
presented hints, to increase the likelihood that behavior will be produced
|
|
relearning method
|
assumes that the less training required to reach the previous level of performance, the less forgetting has occurred.
|
|
savings method
|
a savings of information compared to the original training program
|
|
recognition
|
the participant has only to identify the material previously learned.
|
|
delayed matching to sample
|
the animal is prevented from performing following presentation of the sample.
|
|
extinction method
|
immediate extinction training vs. training following removal from the chamber. if extinction proceeds more rapidly after the retention interval, fogetting has occurred
|
|
gradient degradation
|
forgetting may be measured as a flattening of a generalization gradient. a generalization gradient indicates forgetting.
|
|
variables in forgetting
|
passage of time: the longer the interval between training, the greated the deterioration;
time is not an event, so can't cause other events. DEGREES OF LEARNING, overlearning, PRIOR LEARNING, proactive interference, paired associate learning, SUBSEQUENT LEARNING, retroactive interference, CONTEXT, cue-dependent forgetting, reminiscence, dependent learning |
|
Overlearning
|
practicing even beyond mastery
|
|
fluency
|
number of correct responses per minute while learning
|
|
proactive interference
|
old learning can interfere with recall
|
|
paired associate learning
|
used to study proactive interference; the person learns a list of word pairs, so that when given the first word, can recall the second.
|
|
retroactive interference
|
when what we learn interferes with our ability to recall earlier learning.
|
|
cue-dependent forgetting
|
learning occured with presence of a particular stimuli and that stimuli is now absent.
|
|
reminiscence
|
when performance improves with the passage of time
|
|
state-dependent learning
|
behavior that is learning during a particular physiological state is lost when that state passes.
|
|
foraging
|
the forgetting curve may be less steep when what is learned is important to survival.
|
|
eyewitness testimony
|
dependent on way questions are phrased
|
|
ways of overlearning
|
mnemonics, method of loci, peg word system,
|
|
mnemonic
|
any device for aiding recall, ex: rhyme, acronym
|
|
mnemonic systems
|
method of loci, peg word system
|
|
method of loci
|
apply grocery list to an imaginary walk through a place you frequently visit.
|
|
peg word system
|
one is a bun, two is a shoe
use concret objects as pegs. |
|
context clues
|
study in the classroom
|
|
prompts
|
note on calendar
|
|
synesthesia
|
a synthesis of different senses; sound produces light, color, taste, odor, touch
|
|
limits on learning
|
physical structure: incapable of speaking; learned behavior is not inheritable, dies with the learner
|
|
Lamarck
|
theorized that physical characteristics were acquired adaptations that were passed on from generation to generation.
|
|
heredity
|
plays important role in human learning abilities within and between species.
|
|
environment
|
enriched environments may create intelligence
|
|
neurological damage
|
prenatal exposure to alcohol results in limited learning abilities;
neurotoxin damage nerve tissues; head injury, blow to head or shaking of child; disease and malnutrition: prevent normal neurological development |
|
critical periods
|
stages that are optimum for learning. certain opportunities may occur once in a lifetime.
|
|
imprinting
|
Lorenz (1952) the youngster will become attached to any moving object that happens to pass by, whether another animal of the same species, a mechanical object, or a human being
|
|
Breland (1961)
|
innate tendencies interfere with learning
raccoon not letting go of coin |
|
Garcia and Koelling (1966)
|
taste aversion
|
|
continuum of preparedness
|
an organism comes to a learning situation genetically prepared to learn(quick), unprepared(steady but slow), or contraprepared(slow and irregular)
|
|
Seligman (1970)
|
continuum of preparedness
|