• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/32

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

32 Cards in this Set

  • Front
  • Back

Operant conditioning

Learning theory that identifies the factors responsible for the acquisition and maintenance of complex voluntary behaviors

Operant conditioning (Thorndike)

Learning occurs through the connections between stimuli and responses as a result of trial-and-error.



The behaviors are instrumental in the achieving of a goal (instrumental learning)

The law of effect

Any response followed by satisfying state of affairs (drive reduction?) is likely to be repeated again.



Behaviors resulting in an annoying state of affairs is less likely to occur

BF Skinner

Father of operant conditioning.


Believed that classical conditioning explained automatically elicited behaviors but failed to explain more complex behaviors that are voluntarily emitted.



Operant conditioning is a theory of behavior based on the way behaviors operate on the environment

Reinforcement and punishment

-A reinforcer always increases the probability or intensity of a response occurring.


- A punishment decreases the probability of or prevents a response occurring



Reinforcer:


-Positive refers to the application of a stimulus.


-Negative refers to the withholding or removal of a stimulus.

Positive and negative punishment

Positive punishment: dog chews shoe, hit with newspaper



Negative punishment (response cost): child hits sibling, takes away fav toy.

Operant extinction

Reward withheld from a previously reinforced behavior to eliminate or decrease that behavior.



Usually not immediate but rather gradual reduction in frequency/intensity of a response

Extinction burst

A temporary increase in responding during extinction

Behavioral contrast

When 2 different behaviors are reinforced and then the reinforcement of one behavior is withdrawn in order to extinguish it, the other behavior is likely to increase

Primary unconditioned reinforcers

Inherently desirable and do not acquire reinforcing value through experience

(Generalized) Secondary reinforcers

Acquire reinforcing value through their association with a primary reinforcer.



Ex. A token which is reinforcing because it is associated with a stimulus that is naturally reinforcing

Continuous reinforcement

Reinforcement after every response

Intermittent reinforcement

Reinforcement after a specified number of responses

4 types of intermittent reinforcement schedules

1. Fixed interval



2. Variable interval



3. Fixed ratio



4. Variable ratio

Superstitious behavior and S - R generalization

Accidental, non contingent reinforcement. Reinforcement not tied to any particular behavior



Odd, ritualistic behavior: Behaviors that the organism was engaging in just prior to the non contingent behavior

Discriminated stimuli

Similar, but different, stimuli are differentially reinforced. Organism learns to discriminate between stimuli that are reinforced and similar stimuli that are not reinforced

Escape and avoidance conditioning

Behaviors that are maintained by negative reinforcement (Escape Behavior incr bec the organism's performance allows the organism to escape an undesirable stimulus)



Escape behavior is a two-factor form of learning (organism learns to identify a stimulus that signals the initiation of an aversive stimulus).



If the organism performs the target behavior in the presence of a cue, the organism can escape the negative reinforcer.



Two factors = discrimination learning (cue) and avoidance, or escape learning

Contingency

The reinforcer should only be available when the target Behavior has been performed

Immediacy

The reinforcer should be delivered immediately after the target Behavior. Reinforcement delayed may not reinforce the target Behavior

Schedule of reinforcement

Establishment of a new behavior is most rapid with A continuous reinforcement. Maintenance of the target behavior is maximized with an intermittent schedule of reinforcement

Thinning

The change from a continuous to intermittent reinforcement schedule once the behavior is well-established, then reducing the proportion of reinforcement to Target Behavior

Magnitude of reinforcement

the greater the reinforcer the greater its effectiveness.



Beyond a certain magnitude, satiation may occur (the reinforcer loses its reinforcing value)

Prompts

Verbal or physical reinforcements facilitate the acquisition of the target Behavior

Fading

The gradual removal of the prompt

Shaping

Rewarding behaviors that approximate the target behavior.



Essentially, behaviors come closer and closer to the target behavior

Chaining

Developing a chain/sequence of behaviors in which each subsequent behavior is contingent on what came before

Premack principle

A high probability behavior is used to reinforce a low-probability behavior.



Ex. Video game-playing (high probability behaviour), working on dissertation (low-probability behaviour) - play video game after 1 hour of dissertation work

Differential reinforcement

Combining positive reinforcement with extinction.



Reward 1 of 2 competing responses. Reward one behavior while ignoring the other behavior

Fixed interval

Scalloping. The interval is the same after each reinforcement. The first response is rewarded only after a specified amount of time has elapsed.



Ex: reinforcing a rat with a pellet for the first bar press after 30-second interval

Variable interval

Steady but relatively low level of response. Occurs when a response is rewarded after an unpredictable amount of time has passed.



Ex: delivering a food pellet to a rat after the first bar press following a 1 minute interval, 5 minute, then 3-minute interval

Fixed ratio

Relatively High, steady rate of responding. A response is reinforced only after a specified number of responses.



Ex: delivering a food pellet to a rat after it presses a bar five times

Variable ratio

Highest rate of responding, most resistant to Extinction. When a response is reinforced after an unpredictable number of responses.



Ex: delivering food pellets to a rat after one bar press, again after 4 bar presses, and two bar presses.