• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/70

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

70 Cards in this Set

  • Front
  • Back
Drive reduction theory of reinforcement
Organisms defend a stable state/homeostasis. Stimuli that reduce a drive state are effective reinforcers. Primary and secondary reinforcers, conditioned drive state
Premack Principle theory of reinforcement
The opportunity to perform the higher probability response will serve as a reinforcer for the lower probability response.framed issues of reinforcement in terms of response, not stimuli.
Hard to assign precise numerical value to the porb. of a response
also, likelihood of a response may change unexpectedly
PP doesn’t tell us how reinforcers work
behavioral regulation/bliss point
determine baseline response, limit responding below the baseline
predicts accurately that even a low-prob response can be made a reinforcer
avoids problem of computing response probs in PP, but still doesn’t answer how a reinforcer produces an increase in the porb of the rein. response

*behavioral homeostasis, impose contingency and challenge homeostasis
*minimal deviation model-allocate responses accordingly so that you can get as close to your bliss point/homeostasis as you can
elasticity
highly elastic items very sensitive to price/response requirement, many substitutes, high price, lower income (greater the effect of price)
elastic graph curves downward
inelasticity
highly inelastic items not sensitive to price/response requirement, not many substitutes, low price, greater income (lower the effect of price)
Ex. gas
inelastic graph straight horizontal line
economic model of behavioral regulation
Restrictions on behavior: consumer behavior driven by price of object,and income. organisms will allocate behavior to some optimal point
chosen by organism to maximize something
sort of like hedonism
matching law
relative rate of response will match the relative rate of reinforcement
will match behavior to relative rate of returns
predicts hyperbolic discounting: the further away the returns are, the less we value it
different than what economic theories predict (maximization)
Gordon Foxall's Behavioral Perspective Model - Operant conditioning
types of consequences or outcomes
hedonic reinforcement: you get something that you enjoy, you like outcome so you respond again.
informational reinforcement: status. i.e. ugg boots, BMW
aversive stimuli: spend $, health changes, time spent in acquisition
punishment. Decrease consumer behavior instead of increase
Consumer must do a cost-benefit analysis
Categories of consumer behavior according to Foxall and the BPM
maintenance: things you buy for survival and obligation
accumulation: saving or collecting. Buying things just to have them.
pleasure: consumption of things like popular entertainment
accomplishment: status purchase. showing yourself to be better
maintenance/inelastic-accomplishment/elastic
Behavioral Ecology of Consumption
much consumer behavior is a biobasic behavior and thus subject to evolutionary pressures
search (foraging), choice, consumption, disposition of good
Testing if you have stimulus control
change the features of the stimulus to test
or look at the absence/presence
should see a change in behavior
stimulus generalization
behavior on vertical, stimulus on horizontal
wider curve-generalization (steeper curve, less generalization)
other similar stimulus - due to a spread of the effective training
if can’t detect difference, also might get generalization- due to lack of ability
stimulus discrimination
However, subjects can often be taught to discriminate stimulus and only respond to a specific stimulant.
steep curve, more discrimination
overshadowing
high intensity better for controlling behavior in general
BUT overshadowing: weak tone w/ bright light, no learning to tone because light overshadows (compound CS), also if something in context is really complex compared to the CS
quality-location effect
Type of response required
training dog in right/left discrimination, metronome in front-right leg, buzzar in back-left leg,
group 2 is go/no go
switch positions of metronome and buzzar
location of sound didn’t matter in group 2, but did in group 1
belongingness
visual cues-food, auditory cues-pain avoidance
stimulus element vs. configural cue
different elements may get learned differently by different organisms
configural cue: when compound CS all together, how it’s fit together important
discrimination
can train organisms to have differential responses to different stimuli
discrimination training: training 2 stimuli (1 as S+, the other as S-), you get more discrimination, steeper curve
Better learning with CS+ and - and S+ and - together because more discrimination
expertise result of intradimensional training
discrimination training and role of excitatory and inhibitory conditioning
CS+ vs. CS-
control behavior differently-excitor vs. inhibitor
S(R-O): in the presence of S, R will lead to O
S+ or S^D, discriminative stimuli, excitatory learning
S-, in presence of S, R will not lead to O, inhibitory learning
can train organisms to have differential responses to different stimuli
discrimination training: training 2 stimuli (1 as S+, the other as S-), you get more discrimination, steeper curve
conditioned compensatory responses
drug opposite response tries to protect body from overdose.
UR and CR are opposite
leads to conditioned tolerance
Opposite CR will occur if UR disrupts homeostasis-kind of.
Siegel articles-you did a hw on this!
self-administration cues
SAC= self-administration cues, proprioceptive
tying off of arm reliably predict heroin
physical interoception effects that happen
extinguishing these
allow participants to mimic these cues, just don’t actually administer drug
drug-onset cues
DOC=drug-onset cues, interoceptive
the first sip of beer (reliably predicts 12 beers)
a little bit of the drug
extinguishing these
give ppl tiny bit of drug, then don’t allow them to have any more
role of drug conditioning in drug-seeking (drug-like and drug-opposite responses)
& ways to decrease its influence
see notes!
Extinction
decrease in responding
not forgetting or unlearning or opposite of acquisition
just learning a new relationship
protocol s-r- no o!
frustration/frustrative aggression
What happens when don’t get reinforced? Response to extinction to conditions
high emotional response, frustration/frustrative aggression
frustration burst- a whole bunch of responding at once
but net result of extinction is to decrease responding
Spontaneous recovery
How we know that extinction is not unlearning
delay after extinction training, more acquired responding again
if the acquired response comes back, you did not unlearn it!
reinstatement
acquisition, extinction
if exposed to US, that’s enough to have conditioned responding come back
drug literature uses reinstatement to mean a return of conditioned responding not necessarily in response to exposure to US
renewal
shift testing context cues from extinction training context
if change contextual cues, recovery of acquired responding/performance
huge problem for extinction therapy/desensitization
fear may come back outside therapy context
so try to do the therapy in different contexts...but can’t get to every possible context
so try to do therapy in contexts where you are most likely to associate the stimulus
drug extinction-interoceptive cues especially hard to extinguish?
Extinction is context specific- way more so than acquisition
CS/US devaluation
If extinction leaves intact a CS-US association, then US devaluation will disrupt responding to an extinguished CS, just as it disrupts responding to a CS that has not undergone extinction
overtraining
overtraining actually leads to faster extinction
so certain of rule, that when you’re not reinforced/when rule is violated, you realize it’s a new association/you give it up
partial reinforcement
Partial Reinforcement Resistance to Extinction/ or Part. Rein. Ext. Effect
continuous reinforcement participants will extinguish faster than partial reinforcement participants
partial reinforcement schedule fosters persistence in responding in the face of nonreward
reinforcement magnitude
magnitude of reinforcement effect
extreme version of negative contrast (big to little, decrease in responding)
paradoxical because you originally had a stronger response, yet extinguishes faster
Frustration theory to Partial Reinforcement Resistant to Extinction
learn to expect nonreward and to continue in the face of frustration and nonreward. In anticipation of not being reinforced, you keep responding (even though you’re frustrated). emotional theory
works better with long ITIs because may not remembered when you were rewarded last
Sequential theory to Partial Reinforcement Resistant to Extinction
remember that reinforced trials follow nonreinforcement. Remember bhbb recent responses. Learn that nonrewards preced reward. memory theory.
works better with shorter ITIs
Avoidance trials
giving a signal before the aversive stimulus starts. after one or two occurrences of the punishing stimulus the cue will trigger an avoidance behavior. This kind of learning occurs quickly and is very durable.
escape trials
occurs when the animal learns to perform an operant to terminate an ongoing, aversive stimulus. It is a "get me out of here" or "shut this off" reaction, aimed at escape from pain or annoyance. The behavior that produces escape is negatively reinforced
interdimensional discrimination
discriminating b/w 2 dimensions: white chocolate vs. dark chocolate. male vs female
intradimensional discrimination
5% chocolate vs. 70%, discriminating w/in females. More expertise with intradimensional training!
stimulus equivalence
if say stimulus equivalence, no discrimination b/w them
if organism doesn’t tell you this is diff than this
1. cannot discriminate
2. or learned to not discriminate
stimulus equivalence classes-organism has been trained to think of these things as the same (over time learn A=a). want same responding for all objects in this equivalent class, i.e. sometimes discrimination is bad!
Avoidance theory
*avoid punishment by avoiding stimuli associated with punishment and do something else
but don’t always know what the cues are that the organism is paying attention to.
*Can test avoidance behavior in a shuttle box
Rat has to run to other side of box in order to avoid getting shocked
1 way vs. 2 way avoidance = 1 side shocking vs. 2 sides shocking. 1 way is easier to learn
discriminated avoidance
if there’s a signal for the US
If participant responds when a CS or warning stimulus is presented, the CS turns off and no aversive US.
2 types of trials: response and no-response
free-operant avoidance
No explicit warning stimulus or discrete trials. Can avoid at any time and this response always provides some measure of benefit. EX. changing oil in your car to prevent/avoid engine troubles
R-S interval
In avoidance procedures, interval after avoidance response and before next aversive stimulus presentation. Most net benefit if respond at end of R-S interval.
Use timing cues even though there is no
Problem in analysis of avoidance behavior
how is nothing reinforcing? If i engage in avoidance response, nothing happens. somehow nothing is reinforcing
Two-factor theory of avoidance
2 forms of conditioning used in avoidance learning.
1. Pavlovian conditioning of fear to a stimulus that signals aversive stimulation AND
2. instrumental conditioning of the avoidance response by fear reduction
*fear reduction is negatively reinforcing
problems w/ 2-factor theory
*free-operant avoidance procedure. Warning stimulus required for 2 factor theory
*Many avoidance responses should result in Pavlovian extinction of fear and then extinction of the instrumental avoidance response, but this cycle does not happen.
*After many avoidances, ppl don’t seem fearful
safety-signal hypothesis of avoidance behavior
*positive reinforcement.
*(internal/interoceptive cues) associated with making avoidance response signal safety. Classical conditioning. Safety signal and CER learning can happen at same time! 2 diff ways of thinking about how avoidance learning comes to be.
shock frequency reduction theory
(temporal cues) Will distribute responses at end of interval to reduce frequency of responses? You learn temporal coding, timing. You learn r-s interval, respond at end to reduce frequency. Shock interval = 30 seconds, respond at 29 seconds
Effective punishment
delay, contingency, alternative reinforcers, stimulus control, intensity
consistent, immediate, severe (to sensitize to future punishments), not signaled (if want global decrease in behavior), have alternative ways to earn reinforcers
after school programs
Theories of punishment
Negative Law of Effect
Premack Principle
Conditioned emotional response
Avoidance theory
CER- theory of punishment
Stimulus (context)- elicits emotional response:fear , o=shock, r avoidance response
reduction of fear negatively reinforces you making avoidance response
*Goes with modern 2 process theory
don't see cycles like with modern two process theory
Negative law of effect
*Original law of effect says if O is a satisfying state of affairs, it stamps in the S-R relationship
* If O is Annoying state of affairs, it weakens the S-R relationship. Punishment works cuz s-r relationship is weakened by annoying state of affairs
Premack Principle
*Original: Increase low probability behaviors by making high probability behaviors contingent on them.
*4 Punishment: To decrease responding (as in punishment), make low probability behavior the outcome of high probability behavior.
EX. Decrease movie watching: movies contingent on cleaning room
Working memory
Retention of info needed to respond successfully on one trial or task (in matching-to-sample procedure) but is not useful in responding on the next trial or task. WM is of limited duration
*Variations in sample stimulus and location of correct choice ensure matching-2-sample procedures involve working memory
EX. baking cake, must remember what ingredients you've already included. Won't help you on the next cake
Reference memory
helps with remembering the constant features of the task, longer duration than working memory.
EX. general cooking knowledge about cake pans, ovens, etc.
acquisition, retention, and retrieval
* When testing learning both behav. and memory are important. Can't show learning if don't remember it.
*can look at acquisition, retention, and retrieval - errors in any one of these places can manifest as the same behavior
*problems with acquisition and retention- maybe learning never actually occurred or it did and you just can’t get it out.y
acquisition
Info about stimuli and responses are acquired and encoded in the nervous system. Studies of learning focus on manipulating the acquisition stage
retention
Once encoded, info stored for later use
retieval
info recovered from storage
delayed matching to sample
(specific rule (triangle-i must pick triangle...if never change sample, always use specific rule) vs. general rule- match the sample is general rule, pick triangle is specific rule. Mix it up (as far as samples) to teach general rule)
Trials unique procedure: could never learn specific rule, because there is none (change sample every time)...might be harder for them to learn
trace decay hypothesis
Passive memory
*memory trace forms in mind, degrades
*also, the longer you’re exposed to stimulus, the better trace you’ll get
*support: longer you look at it, better your performance-stronger trace
*however, if you have practice/training with delay, the delay does not have as much impact- argues against trace-delay hypothesis
actually something ACTIVE going on
retrospective vs. prospective coding
Retrospective: think about what you’ve already done, thinking backwards. Memory of previous stimuli/events
Prospective: think about where you’re going, what you need to do, thinking forwards. Remembering what is predicted to occur.
rats will use whichever is closest- switch from retrospective to prospective
spatial memory
just as a rat's spatial memory is needed to learn the location of food at the end of a maze. It is often argued that in both humans and animals, spatial memories are summarized as a cognitive map. Spatial memory has representations within working, short-term and long-term memory.
win-shift
radial arm maze, spatial learning with rats
win-shift procedure: bait certain arms one time and others on the next trial
actually easier for them to learn, works with natural foraging tendencies
belongingness
win-stay
radial arm maze, spatial learning with rats
win-stay procedure: every trial same arms are baited. Harder to learn than win-shift for rats.
retrieval cues
Stimuli previously associated with target info. If poor performance is due to retrieval failure, procedures that facilitate retrieval should facilitate performance.
forgetting vs. retrieval failure
You may not remember because you never encoded or learned the info in the first place. Or you may fail to effective retrieve the info that was successfully encoded or stored.
problems with retrieval : use of retrieval cues helps with this
making environment same as where they learned it S(R-O)
state dependent learning
retroactive interference
Extraneous stimuli can act backward to disrupt memory. Remember too little.
Meet Jane.
Meet Janice.
Can't remember Jane's name.
as you get further in the list, switch from what you’ve done to thinking about future, which gives you least load on your memory
proactive interference
Extraneous stimuli occur b4 target even to disrupt memory. Remembering too much.
EX. You learn about punishment in politics class. This proactively interferes when you go to psych class and try to learn about punishment there.
classical conditioning & marketing/advertising
liked and disliked music (US)
UR liking music
blue and beige pens (CS)
CR liking/disliking pen
taking pen= operant behavior, O= having pen you like
classical and operant conditioning both at work