Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
82 Cards in this Set
- Front
- Back
How did Thorndike's "Puzzle Box" experiments change the way people thought about animal learning?
|
Before the experiments, people believed that animals learned through reasoning. Following the experiment, people were forced to reason that animals were not that intelligent.
|
|
what did Thorndike do? And what was his major finding?
|
Thorndike did the Puzzle Box experiment. He discovered The Law of Effect
|
|
What is the Law of Effect?
|
Behavior is a statement of it's consequences.
|
|
Why is the Law of Effect historically important?
|
Thorndike was the first to show that behaviors is systematically strengthened or weakened by it's consequences
|
|
What is a Skinner Box?
|
Chamber for rats that's designed so that a food magazine could automatically drop a few pellets of food into a tray. The rat would be accustomed to the sound of food dropping and a lever would then be made that the rat would have to press.
|
|
Why is the Skinner Box useful for Learning and Memory?
|
The Skinner Box is a convenient way to provide consequences to the on-going behavior of an organism, without having to interrupt it by handling.
|
|
What is Operant Learning?
|
When consequences strengthen or weaken a behavior.
|
|
What are the two main types of Operant Learning?
|
Reinforcement & Punishment
|
|
Define Reinforcement
|
An increase in the strength of behavior due to it's consequence.
|
|
Positive Reinforcement
|
Behavior is followed by the appearance of or increase in the intensity of a stimulus.
|
|
Example of Positive Reinforcement
|
Cookies and words of encouragement and prizes. "Reward Learning".
Daughter gets and A on a test. I give her a cookie. She wants to get an A again so she gets another cookie. Cookie = + Reinforcer |
|
Negative Reinforcement
|
When a behavior is followed by removal or reduction in intensity of a stimulus.
|
|
Examples of Negative Reinforcement
|
Daughter gets and A on a test, I tell her she doesn't have to do the dishes since she got an A. She wants to get an A again so she doesn't have to do dishes. Dishes = - Reinforcer
|
|
Reward vs. Reinforcer
|
Reward is a non-technical term that should be avoided. Reinforcer is a technical term that implies a functional relationship between a behavior and a consequence.
|
|
Aversive Stimulation & Negative Reinforcement
|
Both are removing a stimulus to increase a behaviors.
|
|
What are 2 types of negative reinforcement?
|
Escape & Avoidance
|
|
Escape - Negative Reinforcement
|
Responding removes an aversive stimulus. Ex: Turn down blaring car stereo/Punching Will in the face.
|
|
Avoidance - Negative Reinforcement
|
Responding presenting onset of an aversive stimulus. Ex: Turn down volume knob before turning on car/avoiding where Will is.
|
|
What is typically meant by the "strength" of a behavior?
|
Typically we are referring to increases in rate of behavior, however, we could also be referring to persistence of a behavior once reinforcement is withheld or the behavior is otherwise disruptive.
|
|
2 learning procedures
|
Discrete trials & Free-Operant
|
|
Discrete Trial
|
Behavior ends a trial. Ex: Puzzle fucking box
|
|
Free-Operant Learning
|
Behavior is ongoing and can be repeated without disruption. Ex: Skinner's Box
|
|
Operant vs. Pavlovian
|
Pavlovian Conditioning involves stimulus stimulus learning, Operant conditioning involves behavior stimulus learning.
|
|
How useful is the voluntary-involuntary distinction?
|
Not very... >.>
Because it's hard to distinguish between the two. Sometimes behavior that seems involuntary(pavlovian) can be in fact voluntary (operant). The two can even overlap |
|
3 term contingency
|
Involves Stimulus-Behavior-Stimulus Learning :)
|
|
Primary Reinforcers
|
Are those that are naturally or innately reinforcing. Ex: Food, water, sex, warmth
|
|
Secondary Reinforcers
|
Are those that are effective due to association with other reinforcers. Ex: Praise, smiles, money. AKA Conditioned reinforcers
|
|
Disadvantages to Primary Reinforcers.
|
Tend to lose there effectiveness quickly. Ex: You can only eat so much.
|
|
Advantages of Secondary Reinforcers
|
Don't lose there effectiveness as quickly as Primary. Easier to deliver immediately. Don't interrupt ongoing behavior. Ex: Money is more useful than warmth because u can buy different forms of warmth.
|
|
What is the contrived distinction?
|
Contrived is explecitly arranged by someone to modify behavior. Ex: Food pellets for rats, access to grain fro pigeon, praise for baby.
|
|
What is Natural Distinction?
|
Are events that follow automatically from a behaviors. Ex: Light switch = Light, arriving at a destination after navigating there. They are important because they are expected and immediate.
|
|
What are 3 ways reinforcers shape new behavior.
|
Shaping, chaining, Direct Reinforcement
|
|
What is Shaping?
|
Reinforcement of successive approximations of an end behaviors. Ex: Parent can train more misbehavior by waiting longer to give into demands.
|
|
What is Chaining?
|
The process of training an interrelated series of acts or behavior chain. There are 2 types, forward and backward.
|
|
What is Forward Chaining?
|
The tasks are trained in the order that they will be in when the behavior is complete. Ex: Tying your shoes
|
|
What is Backward Chaining?
|
The tasks are trained in reverse order; the last part is trained first followed by adding the next to last act and so on. Ex: Marching Band steps.
|
|
What is meant by Contingency? And how does it affect Operant Learning?
|
Variable affecting reinforcement it refers to an if/then relationship between 2 things. In the case of Operant Learning, contingency refers to the degree to which the behaviors produces the reinforcer.
|
|
How does Contiguity affect Operant Learning?
|
Contiguity means the closeness of 2 events in time. In the case of Operant Learning, contiguity refers to the delay between the behaviors and the consequence.
|
|
Reinforcers characteristics (3)
|
1) Larger Reinforcers typically produce faster learning and higher response rates
2)Continuing to increase reinforcer magnitude will not result in continued increase in responding 3)Qualitative differences can play a role as well. Ex: Rats prefer icecream to water. |
|
Motivating (Or establishing) operations
|
Any situational arrangement that increases the effectiveness of a reinforcer.
|
|
What is Extinction as it relates to Operant Learning?
|
Allow the behavior to occur but cease to enforce it: The Behavior Stimulus ceases to exist.
|
|
What is Extinction Burst?
|
When previously reinforced behavior is no longer reinforced, there is often an initial increase in rate or intensity of the behavior before a decline. Problems - it can look like the behavior targeted for reduction is gettin' worse.
|
|
What is Spontaneous Recovery?
|
Behavior can occur even after thousands of un-reinforced responses.
|
|
What is Resurgence?
|
The reappearance of other previously reinforced behavior during extinction. Ex: counting on fingers
|
|
What is Drive Reduction Theory?
|
Drives are motivational states the produce tension in the organism when present. Ex: Hunger, horny, sleepy.
Reinforcer is reinforcing because it reduces 1 or more drives. |
|
Weakness of the Drive Reduction Theory
|
Food in reinforcing because it reduces food drive. How do u know it reduces food drive? Because it is reinforcing. Hmmmm.... It's circular reasoning.
|
|
Relative Value Theory
|
Theory of reinforcement that considers reinforcers to be behaviors rather than stimuli and that attributes are reinforcers effectiveness to it's probability relative to other behaviors.
|
|
Problems with Relative Value Theory
|
1) Does not explain why one behavior is more or less reinforcing than another.
2) It's hard to explain things like praise as a reinforcer. |
|
Response Deprivation Theory
|
The theory of reinforcement that says a behavior is reinforcing to the extent that the organism has been deprived of forming that behavior.
|
|
Problems with Response Deprivation Theory
|
It's hard to explain things like praise as a reinforcer.
|
|
What is the Premack principle? And what theory does it belong to?
|
The Premack Principle belongs to the Relative Value Theory. Premack Principle is the observation that high probability behavior reinforces low probability behavior.
|
|
In what way does avoidance learning present a theoretical conundrum?
|
When an organism learns to avoid an aversive stimulus, the aversive stimulus no longer occurs. So then, what maintains the avoidance behavior? There are 2 major theories, the 2-process theory and the 1-process theory.
|
|
What is 2-Process theory?
|
2 kinds of learning are involved in avoidance learning. Operant and Pavlovian.
|
|
The problems with the 2-process theory?
|
Avoidance behavior itself should start to extinguish but it doesn't.
|
|
What is 1-Process theory?
|
Only Operant Learning is involved in avoidance; avoidance can be understood as reducing the overall frequency of averse stimulation.
|
|
What is a Schedule Reinforcer?
|
Different rules by which reinforcers are delivered
|
|
What is a Schedule Effect?
|
Changes in behavior due to schedules of reinforcement
|
|
What is Continuous Reinforcement?
|
The simplest schedule of reinforcement. Where every instance of the behavior is reinforced.
|
|
What good is Continuous Reinforcement for?
|
Best schedule for training new behavior.
|
|
What is an Intermittent (or partial) reinforcement schedule?
|
Any reinforcement schedule that doesn't reinforce every instance of a behavior.
|
|
Name and describe each of the 4 main simple schedules and give examples.
|
Fixed Ratio (Every 5 times u hit the lever, u get a treat), Variable Ratio (Golf or gambling), Fixed Interval (Morphine drip), Variable Interval (Predator waiting for prey)
|
|
Fixed Ratio (Effect on behavior)
|
High rate of response is typical. Post-reinforcement pause is usually observed in animals
|
|
Variable Ratio (Effect on behavior)
|
Very high rate of response is typical. Post-reinforcement pause is sometimes observed in animals
|
|
Fixed Interval (Effect on behavior)
|
Between reinforcers, changing rate of response from low to high is typical
|
|
Variable Interval (Effect on behavior)
|
Moderate, steady rate of response is typical. Post-reinforcement pausing does not occur.
|
|
Fixed Ratio (What it looks like on graph)
|
Overall pattern is then pause-run-pause-run, which produces a step-like cumulative record
|
|
Variable Ratio (What it looks like on graph)
|
Overall pattern is generally run-run-run, which produces a steep, straight cumulative record.
|
|
Fixed Interval (What it looks like on graph)
|
Overall pattern of response is pause-run-pause-run, which averages into the familiar scallop shape
|
|
Variable Interval (What it looks like on graph)
|
More steep, few pauses
|
|
What is a Duration Schedule?
|
Reinforcer is contingent upon doing something continuously for a specific length of time. Can be fixed or variable. Ex: Praise contingent upon 15 minutes of continuous excercise.
|
|
What is a Rate Schedule?
|
Reinforcer is contingent upon emitting behavior either quickly or slowly enough.
|
|
Rate Schedule - Differential Reinforcement of Low Rate (DRL)
|
Reinforcer is contingent upon emitting an inter-response time (IRT) above t s.
|
|
Rate Schedule - Differential Reinforcement of High Rate (DRH)
|
Reinforcer is contingent upon emitting an IRT below t s OR emitting a certain number of responses within a specified time period.
|
|
Rate Schedule - Differential Reinforcement of paced responding schedule (DRP)
|
Reinforcing only IRTs with minimum and maximum constraints
|
|
What is a Non-contingent Reinforcement (NCR)
|
Reinforcer is delivered regardless of behavior.
|
|
Why do some object the term "Non-contingent Reinforcement"?
|
Some object to even referring to this as a schedule of "Reinforcement" because there is no contingency between behavior and reinforcer.
|
|
What is a Progressive Schedule?
|
Some parameter is increased in a regular manner.
|
|
Partial Reinforcement Extinction Effect (PREE)
|
Animals trained to complete a particular task using partial reinforcement (as opposed to Continuous reinforcement) show resistance to extinction: that is they persist in responding when the reinforcement is no longer being delivered for greater lengths of time than animals that were trained with continuous reinforcement.
|
|
What are four theories that attempt to explain Partial Reinforcement Extinction Effect (PREE)?
|
1) Discrimination Hypothesis
2) Frustration Hypothesis 3) Sequential Hypothesis 4) Response Unit - Hypothesis |
|
What is the Discrimination Hypothesis of PREE?
|
Harder to distinguish the transition from partial reinforcement to extinction than from continuous to extinction.
|
|
What is the Frustration Hypothesis of PREE?
|
Extinction engenders frustration, which is relieved by not responding.
|
|
What is the Sequential Hypothesis of PREE?
|
During continuous reinforcement, the reinforcer is a cue to continue responding.
|