Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
29 Cards in this Set
- Front
- Back
internal validity
|
elimination of plausible alternative causes in an experiment i.e. the independent variable caused the result
|
|
independent groups designs
|
each group of subjects participates in a different condition of the independent variable
|
|
random assignment
|
used to form comparable groups priot to implementing the independent variable. balances/averages individual differences
|
|
random groups design
|
when a random assignment to conditions is used and an observation of each group is made.
|
|
block randomization
|
assign subjects to conditions one block at a time, each block consists of a random arrangement of the five conditions. controls for time related variables and individual differences
|
|
threats to internal validity
|
uncontrolled vaibales that may be a potential alternative explanation for a study's findings
|
|
mechanical subject loss
|
occurs when a subject fails to complete the experiment because of an equipment failure i.e. a computer crashes
|
|
selective subject loss
|
occurs when subj. are lost differentially across the conditions of the experiment, when some characteristic of the subj. is responsible for the loss and when this subj. characteristic is related to the depended variable used to assess the outcome of the study.
|
|
experimenter effects
|
potential biases that arise due to the ecpectationsof the experimenter
|
|
placebo control group
|
a control group that is given a substances that looks like a drug but is inactive
|
|
mechanical subject loss
|
occurs when a subject fails to complete the experiment because of an equipment failure i.e. a computer crashes
|
|
selective subject loss
|
occurs when subj. are lost differentially across the conditions of the experiment, when some characteristic of the subj. is responsible for the loss and when this subj. characteristic is related to the depended variable used to assess the outcome of the study.
|
|
experimenter effects
|
potential biases that arise due to the ecpectationsof the experimenter
|
|
placebo control group
|
a control group that is given a substances that looks like a drug but is inactive
|
|
double-blind procedure
|
both the participant and the observer are blind to what treatment is being administered
|
|
replication
|
repeating the procedures used in a particular experiment in order to determine whether the same results will be obtained a second time
|
|
effect size
|
how large an effect the independent variable has on the dependent variable. Cohen's d. independent of sample size.
|
|
meta-analysis
|
a statistical technique used to summarize the effect sizes from several independent experiments investigating the same independent or dependent variable
|
|
null hypothesis significance testing (NHST)
|
decides whether an independent variable has produced an effect on an experiment i.e. f- or t-test
|
|
statistically significant
|
outcome is one that has only a small likelihood of occuring if the null hypothesis were true
|
|
confidence intervals
|
width of interval tells how precise our estimate is, the more narrow the better. if intervals for two conditions do not overlap then population means differ, when intervals do overlap slightly then there is uncertainty, if intervals overlap where one mean is within another groups interval than pop. means do not differ
|
|
matched groups design
|
researcher makes equivalent groups by matching subjects
|
|
individual differences variables (subj. variables)
|
characteristic or trait that varies across individuals
|
|
natural groups designs
|
experiments involving independent variables whose levels are selected - like individual differences variables i.e. cannot randomly assign people to divorce
|
|
repeated measures designs
|
subj. serve as their own controls bc they participate in both the control and the experimental conditions
|
|
sensitivity
|
the ability to detect the effect of the independent variable even if the effect is a small one
|
|
practice effect
|
the changes participants undergo with repeated testing in the repeated measures designs
|
|
counterbalancing
|
balancing practice effects in the repeated measures design
|
|
differential transfer
|
arises when performance in one condition differs depending on the condition that precedes it
|