• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/105

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

105 Cards in this Set

  • Front
  • Back
  • 3rd side (hint)

Set up a testing procedure for blocking.

Blocking. Control


S1| A+ | --


S2| AB+ | AB+


Test| B- | B-

Set up a testing procedure for latent inhibition

LI. | control


S1|A- | --


S2|A+ | A+

What is the testing procedure for negative patterning

A+/B+/AB-

What is instrumental operant conditioning

Learning the consequences of your own actions

What are the conditioned stimulus processing models?

Mackintosh - associability of the conditioned stimulus is important


If the CS is the best predictor (compared to others CS) of the unconditioned stimulus then Associability of the Cs will be high


if the CS is not the best predictor of the US, the associability to the Cs Will Be Low



Pierce hall - salience and Novelty in CS is important. The associability of a stimulus will be high when it has been followed by a US that is unexpected on the proceeding trial but its associability will be low when it has been followed by US that is expected.

What is one fundamental in instrumental conditioning?

Requires trial and error and multiple trials

What is the law of effect?

If a response to a stimulus is followed by satisfying event (reinforcement), the response is strengthened.


if a response to a stimulus is followed by an annoying event ( punishment) the response is weakened.


Involves SR learning

Explains compulsive Behavior like biting one's Nails snacking or smoking cigarettes

What are the procedures to study instrumental conditioning?

Discrete trail procedures - instrumental response is performed only once. (Eg, Maze experiments) [looking at speed and latency]


Free operant procedures - instrumental response can be performed in a continuous manner. (Many times over again)[looking at operant response]

what are necessary training component used in instrumental conditioning?

Magazine training- training an animal to eat food out of a food well, in a Skinner box


Shaping- reinforcements of successive approximation

If all responses are followed by an outcome, what are the procedures instrumental conditioning uses?

Appetitive stimulus : a pleasant outcome


Aversive stimulus: an unpleasant outcome


Reinforcement: results in an increase in desired Behavior


Punishment: results in a decrease in desired Behavior

Examples of a response producing outcome? (Positive contingency)

Positive reinforcement: is a response produces an appetitive stimulus.


Positive punishment: response produces an aversive stimulus

examples of response preventing or eliminating outcome? (neg contingency )

Omission training or negative punishment: responses eliminate or prevent the occurrence of an appetitive stimulus.



Negative reinforcement: response eliminate or prevent the occurrence of an aversive stimulus.

What are the fundamental elements of instrumental conditioning?

Instrumental response


outcome


contingency between the response and outcome

What are the types of instrumental responses?

Serotype responses: the same response is required to get the outcomes over a number of trials



Variable responses: different responses in each trial are necessary to get the outcome

What is belongingness and the relevance of it?

Certain responses belong with the reinforcer because of the animal's evolutionary history.

What is instinctive drift?

Response And other instinctive behaviors with drift in and interfere with performance of the operant response

What variables determine the effectiveness of positive reinforcement?

Quality and quantity and it can be seen on a progressive ratio schedule.

What is the crispie effect?

Positive contrast: low-to-high (reward ) results in increased instrumental response


Negative contrast: high to low results in a decrease in instrumental response.

What is contiguity?

Time between the conditioned stimulus and the unconditioned stimulus

The researchers that tested behavioral variability and stereotyping and what that their test show?

Paige and neuringer the variable group showed much greater response variability and the control group showed stereotyped responses

What is response reinforcer relation?

Instrumental behavior that is produced is controlled by its consequences


two Parts; temporal relation refers to the time between a response and the reinforcer, special case is temporal contiguity refers to the delivery of a reinforcer immediately after the response


second part; is casual relationship or response reinforcer contingency refers to the extent to which instrumental response is necessary and sufficient to produce reinforcer.

Why is responding so sensitive to delay of reinforcement?

Delay makes it difficult to determine which response is reinforced.

How do you overcome the delay in reinforcement?

1)Provide a secondary reinforcer :a conditioned stimulus that was previously associated with the reinforcer. Primary RF- biological; satisfies a need. 2) provide a marker that Target the instrumental response in some way to make distinguishable from the other activities of an organism;marking procedure (turning on a light) * can block learning if we delay marking procedure closer to outcome.


What was concluded or learned from the response reinforcer relation?

The studies of delayed reinforcement show that a perfect casual relation between the response and outcome is not sufficient to produce strong instrumental responding


led to the conclusion that response reinforcer contiguity rather than contingency to use the critical Factor to produce instrumental learning. However disproved by skinner and showed that conttingency is also important.

What was also learned from the response reinforcer relation?

Contingency is also important


effects of the controllability


A strong contingency, whether the reinforcer occurs depends on whether the instrumental response has occurred.

What is Yoked?

Some aspect of training is linked to another group in the experiment

What is the Learned helplessness effect and how was it tested?

Seilgmann had three groups which included a group that had an inescapable shock and the prior experience to the inescapable shock affected their escape avoidance learning which concluded animals can perceive contingency between their behavior and the reinforcement there's motivational lost in the subject generalized expectations and will continue to be independent of Its Behavior (lack of control)

What challenged the learned helplessness hypothesis?

Learning deficit is not due to lack of control as suggested by the learned helplessness Theory but rather is due to lack of predictability

What do we conclude with continuity and contingency?

Casual relation or contingency ensure that the reinforcer is delivered only the occurance of the specified instrumental response. The contiguity relation ensures that other activities do not intrude between the specified response and the reinforcer to interfere with conditioning the target response.

What is the schedules of reinforcement?

A rule that determines which occurrence of a response is followed by the reinforcer.

What is the simplest schedule of renforcement?

Reinforcement depends only on the number of responses the organism has to perform. Requires Meerly counting the number of responses that have occurred in delivering the reinforcer each time required number is reached more common to have a partial reinforcement procedure if not intermittent reinforcement is used.

What are the four other types of schedules of reinforcement?

Fixed ratio (crs) - subject to reward after a set of responses


Fixed interval: response is reinforced only after a certain time has elapsed


Variable ratio: response is reinforced after a average number of responses


Variable interval: response is reinforced after an average amount of time has elapsed

What are the comparisons of ratio and interval schedules?

Fr and fI have a post reinforcement pause and they produce High rates of responding just before the delivery of the next reinforcer.


Vi & vr maintain steady rates of responding without predictable pauses.



VR produces higher response rates than VI schedules


Ratio schedule produces higher response rates and interval schedules

What is cumulative recording and its signs?

Continuously counts for the responses and animal makes


Horizontal line means no response


Is measure is greater than 0 animals responding


slope is the rate of responding


What are the types of patterns the schedules of reinforcement have?

FR staircase patterns


FI scalloped pattern


VR and VI constant responding because no predictability no reinforcements pauses.

What is limited hold?

Kind of restriction on how long a reinforcer remains available it can be added to either FI or VI schedules.

What did Reynolds find when he compared responses on VI & VR schedule?

1 pigeon was reinforced on a variable ratio schedule the other pigeon on a variable interval ratio schedule the frequency of reinforcement is virtually identical for the two animals


Differences in reinforcement rate do not account for differences in response rate.

What is another possible reason for higher response rates for variable ratio than variable interval?

On a variable ratio schedule a certain number of responses must be made to obtain each reward.


On a variable interval schedule only one response must be made to obtain each reward.

What is inter response time?

Time between each response

So why do ratio schedules produce higher rates of responding than interval schedules?

the inter response time between responses

How do IRT's differ?

The subject is reinforced for a response that occurred shortly after the preceding one than a short IRT is reinforce and short IRT become more likely in the future,.


If the subject is reinforced for a long response that ends a long IRT than a long IRT is reinforced and long IRT's become more likely in the future



ratio schedules favor short IRTs


Interval schedules favor long IRTs

what are response rate schedules?

We can have a procedure that requires that a subject to respond at a particular rate to get reinforced

What are response rate schedules of reinforcement?

DRL - differential reinforcement of low rates of responding , reinforcing long IRT's.


DRH - differential reinforcement of high rates of responding, reinforces short IRT's .

What is a concurrent schedule of reinforcement and what is its purpose?

Concurrent schedules allow for continuous measurement of choice because the organism is free to change back and forth between the response alteratives at any time.


The purpose is to see the distribution of Its Behavior between the two response alternatives.

Animals and people try to maximize the number of reinforcements that they receive what is the exception?

When to FR schedules are used.

What is matching law?

The relative frequency of behavior matches the relative frequency of reinforcement.



B1/B1+B2=R1/R1+R2



B= number of responses


R= rate of reinforcement

What are the problems with matching?

Overmatching higher rate of responding for the better of two schedules than the matching law predicts


usually occurs when it is costly to switch to less preferred response alternative. (Two levers far apart)



Under matching occurs when the subjects respond less than predicted on the advantageous schedule


Relative vs. Absolute reward magnitude


1 pellet vs 3 pellets ( matching law predicts the same)



Response bias individual rights for pigeons may have an unconditioned preference of one choice



Ratio vs interval animals do not match when given concurrent ratio schedule they choose the best one

What are the mechanisms or theories of matching?

Molecular maximizing animals choose whichever response alternative is more likely to be reinforced at that time.



Molar maximizing animals distribute the responses among various alternatives so as to maximize the amount of reinforcement they learned over the long runMelioration animals are continuously attempting to better their current chances of receiving award by switching to the other choice looks at the local rate of reinforcement rather than the overall rate of reinforcement


Melioration animals are continuously attempting to better their current chances of receiving award by switching to the other choice looks at the local rate of reinforcement rather than the overall rate of reinforcement

What are concurrent chain schedules?



Involves two stages choice link and terminal linkThis type of schedule involves Choice with commitment.


This type of schedule involves Choice with commitment.


This type of schedule involves Choice with commitment.

Why do people prefer immediate rather than delayed Rewards?

Value/Delay discounting refers to the concept the value of a reinforcer declines as a function of how long you have to wait to obtain it.

What are the theories for self-control and impulsivity?

Concurrent change schedules have been used to study self control success has been found when value discounting is enforced. (Delayed reinforcement )

What are progressive ratio schedules used for?

To study behavioral economics

reinforcement of successive approximations is a way to ___ Behavior

Shape

Operant conditioning is concerned with how our Behavior can be modified by learning about the ___ of our actions

Outcomes

When the exact same response is made over and over again even if there are other responses possible that behavior is said to become..

Stereotyped

true or false: positive contrast effect is a decrease in instrumental Behavior due to an increase in the amount of reinforcer that is given

False

Providing a secondary reinforcer and the use of a marking procedure is used for what?

To overcome a delay in the outcome

According to the Learned helplessness effect contingency refers to how much ____ the animal has over the outcome.

Control

What effect causes motivational loss and the generalized expectancy that the outcome will continue to be independent of the animal's behavior

Learned helplessness

Which two schedules of reinforcement causes the most consistent levels of responding no (post reinforcement pauses)

VR&VR

Which schedule of reinforcement is being used when the animal received reinforcement after a fixed number of responses?

FR

To overcome a delay between when the animal makes a response and when the outcome is delivered to you use..

A secondary reinforcer

T or F:Adding a feedback cue that predicts the termination of a shock will enhance the learned helplessness effect

False

Schedules of reinforcement are essentially partial reinforcement schedules. T or F

True

In experimental designs in which members of an experimental group and a control group are paired the control group members recieve the same stimuli, renforcement or punishment as the experimental group members but without the possibility influencing these effects through their own Behavior. The control group is sometimes called a ___ group

Yoked

Define the schedules of reinforcement.

A scallop pattern of responding is indicative of what schedule of reinforcement

FI

For interval schedules the inter response times are ___ than ratio schedules

Longer

T or F: DRL schedules have short IRTs

False

According to temporal discounting rather than the actual value of the reward we make our decisions based on their ____ value of the reward at the time when we have to make a decision

Percieved

What is the associative structure of instrumental conditioning?

The way we explain classical conditioning is very cognitive following learning the conditioned stimulus activates mental representation the US which then activates the CR


Similar to instrumental conditioning animals process information with respect to stimuli & to their own behaviors

What are the components of an associative structure?

Stimulus (Sd): Every Instrumental behavior is performed in an environment with specific stimuli (turn key to turn on car)


Response (R)


Outcome (O)

What was determined by the S-R Association?

Thorndikes view - establishment develops between the response and the contextual stimuli that are present when the response was reinforced.


Current view- S-R associations represent habits of Habitual response is an automatic reaction to a stimulus context in which the goal was previously obtained. Devaluation will have no effect on responses.


Ex. People who take drugs even if it is no longer pleasurable explains relapse

What is determined by the S-O Association?

Expectancy of reward


anticipated learning = pavolonian conditioning


Motivates instrumental responding

How does the S-O Association motivate instrumental Behavior?

Two process Theory Pavlovian process and instrumental process

What is the two process Theory say about S-O Association?

S-O activate an emotional state positive outcome in a appetitive stimulus negative outcome is an aversive stimulus


This emotional state (which creates motivation) Summates with the appetitive motivation that is involved in making the instrumental response.

What is the pavlovian instrumental transfer or PIT? what does it signify?

This showed that pavlovian S-O motivates instrumental Behavior. In an experimental design Phase 1(lever press-->food) is constructed of instrumental conditioning


Phase 2 (light--> food) pavlovian conditioning


transfer test (lever press --->food; tone vs no tone) presented pavlovian CS during performance of instrumental response


If pavlovian motivates instrumental Behavior then the rate of lever pressing should increase when the tone CS is presented during the transfer test. This was found.


Results of the transfer test is determined by the expectancies for specific rewards rather than a general positive emotional state

Make an example of a demonstration of S-O.

What is the R-O Association to instrumental conditioning?

Usually it is goal directed ; a response produces outcome an example of this is maze learning and an easy way to show that an animal has developed an R-O situation is to devalue the outcome it should decrease responding for the outcome.


It is most common similar to S-R & easiest to show

Demonstrate an example of R-O.

What are other theories of reinforcement?

Hulls drives reduction- body wants to maintain homeostasis when a need occurs this upsets homeostasis anything that can restore the body back to homeostasis can be a reinforcer



Premack principle- all responses should be thought as reinforcers any response that occurs with a fairly high frequency can be used to reinforce a response that occurs with a relatively lower frequency.


When behavioral preferences change the value of the reinforcer must be evaluated



Response deprivation hypothesis- William Timberlake- the proportional distribution of activities constitutes an equilibrium

post reinforcement pauses are present in which two schedules of reinforcement?

FI&FR

The difference in rate of responding between ratio and interval schedules is a result of differing?

IRT -inter response times

Used in behavioral economics these schedules are used to determine at which price point people will stop purchasing an item.

Progressive ratio

If given a direct Choice an animal will choose the more immediate reward rather than the delayed or even if delayed reward is larger. T or F

T

Match the terms : S-O, S-R, R-O; Habit, Pavlovian, Goal Directed.

S-O= Pavolonian


S-R=Habit


R-O=Goal Directed

To test for presence of associations which two S-O, R-O, S-R , use the devaluation procedure?

R-O& S-O

Is the animal has developed an S-R Association devaluation of the outcome should result in a decrease in responding. T or F

False

Devaluation procedure should show a -- in responding for R-O

DECREASE

PIT is used to demonstrate how S-R motivates instrumental responding. T or F

False

Which two behavioral regulation theories of reinforcements assumes the body wants to maintain homeostasis /equilibrium

Hulls&Timberlakes

If I enjoyed playing the guitar more than taking a shower then according to Premacks theory of reinforcement what should make the best reinforcer?

Playing guitar

According to the two process theory of instrumental responding the S-O association activates a ___ that some of the goal Directed movement.

Emotional state

S-S theory of how we get conditioned responding suggest that after learning the CS activates a mental representation of the US. T or F

True

The Habit Association only happens after extended training. T or F

True

CS processing models differ from us processing models in what way.

Salience of the CS chances

Which phenomenon poses a problem for Rescola & Wagner?

Latent Inhibition

Discrete trial procedures require multiple instrumental responses true or false

False

Squirting your cat with water for eating the plants is an example of what

Positive punishment

Taking away your cell phone for failing an exam is an example of

Omission training or negative punishment

To leave the building when the fire alarm rings is an example of what

Negative reinforcement

Do we test for stimulus stimulus or stimulus response learning?

Devaluation experiments of the unconditioned stimulus see what happens to the conditioned response


Is stimulus stimulus learning the conditioned response will reduce


If stimulus response learning the conditioned response will remain the same

Give an example of unconditioned stimulus processing model

Riscorla and Wagner

Set up for Conditioned Inhibition

Inhibition A+/AB-