Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
113 Cards in this Set
- Front
- Back
sampling distribution
|
distribution of all possible values of a statistic (a picture of the null hypothesis in hypothesis testing)
|
|
statistic
|
a value summarizing a measurable characteristic of a sample (mean of sample, SD of sample, etc.)
|
|
parameter
|
a number that represents the characteristics of a population (mean of population, SD of population, etc.)
|
|
sampling distribution of the mean
|
distribution of all possible means of samples of a given size taken from a population (tends to be normal in shape no matter what the shape is of the original distribution, especially for n>30)
|
|
binomial distribution
|
sampling distribution for events that have two equally likely possibilities (ex: coin flip)
|
|
sampling error
|
amount of error that exists between a sample statistic and a population parameter
|
|
standard error of the mean
|
standard deviation of the means in the sampling distribution of the mean (estimate of the amount of variability in the expected sample means in the sampline distribution of means)
|
|
null hypothesis (Ho)
|
the assumption that observed data reflect only what would be expected from the sampling distribution (no effect of treatment, no relationship b/w variables, no difference b/w means, etc.)
|
|
research (alternative) hypothesis (H1)
|
a prediction regarding the relationship b/w or among two or more variables
|
|
observed value
|
value of inferential test statistic based on observed data
|
|
p-value (probablility value)
|
probability of obtaining an observed value (test statistic) as extreme or more extreme if the null is true
|
|
critical value
|
value of test statistic associated with alpha level (defines criticial region)
|
|
critical region
|
region that identifies values of the test statistic associated with statistically significant outcomes (if the observed value falls in critical region, the null is rejected; if not, the null is not rejected)
|
|
alpha level (significance level)
|
probability of obtaining an observed value in the critical region even if the null were true
|
|
statistically significant
|
the conclusion to reject the null hypothesis, made when the p-value is smaller than alpha
|
|
statistically nonsignificant
|
the conclusion to not reject the null hypothesis made when the p-value is great than alpha
|
|
effect size
|
a statistic that indexes the size of a relationship
|
|
one-sided (one-tailed) p-values
|
p-values that consider only the likelihood that a relationship occurs in the predicted direction
|
|
two-sided (two-tailed) p-values
|
p values that consider the likelihood that a relationship can occur either in the expected or in the unexpected direction (more commonly used)
|
|
steps in hypothesis testing
|
1. develop research hypothesis, 2. set alpha (usually a=.05), 3. collect data, 4. calculate test statistic (observed value) and corresponding p-value, 5. compare p-value to alpha
|
|
p < a (p<.05)
|
reject null
|
|
p > a (p>.05)
|
fail to reject null
|
|
type I error
|
("false alarm") occurs when we reject a true null (conclude a coil is unfair when it is fair) probability=a
|
|
type II error
|
("miss") occurs when we fail to reject a false null (conclude a coin is fair wen it is unfair) probability=b
|
|
correct decision I
|
occurs when we fail to reject a true null (conclude a coin is fair when it is fair) considered a "boring" result. probability=1-a
|
|
correct decision II
|
("hit") occurs when we reject a false null (conclude a coin is unfair when it is unfair) *known as POWER (the ability of a test to fnd a difference when there really is one) probability=1-b
|
|
relationship between a and b
|
when a is lower (probability of making a type I error is lower) then b is higher (probability of making a type II error is higher)
|
|
trade off between type I and type II errors
|
most researchers believe it is worse to make a type I error than a type II error so alpha is ALMOST ALWAYS set lower than beta
|
|
t-test (for independent samples)
|
population mean and SD is unknown
|
|
statistical significance
|
effect size x sample size
|
|
sampling distribution
|
distribution of all possible values of a statistic (a picture of the null hypothesis in hypothesis testing)
|
|
statistic
|
a value summarizing a measurable characteristic of a sample (mean of sample, SD of sample, etc.)
|
|
parameter
|
a number that represents the characteristics of a population (mean of population, SD of population, etc.)
|
|
sampling distribution of the mean
|
distribution of all possible means of samples of a given size taken from a population (tends to be normal in shape no matter what the shape is of the original distribution, especially for n>30)
|
|
binomial distribution
|
sampling distribution for events that have two equally likely possibilities (ex: coin flip)
|
|
sampling error
|
amount of error that exists between a sample statistic and a population parameter
|
|
standard error of the mean
|
standard deviation of the means in the sampling distribution of the mean (estimate of the amount of variability in the expected sample means in the sampline distribution of means)
|
|
null hypothesis (Ho)
|
the assumption that observed data reflect only what would be expected from the sampling distribution (no effect of treatment, no relationship b/w variables, no difference b/w means, etc.)
|
|
research (alternative) hypothesis (H1)
|
a prediction regarding the relationship b/w or among two or more variables
|
|
observed value
|
value of inferential test statistic based on observed data
|
|
p-value (probablility value)
|
probability of obtaining an observed value (test statistic) as extreme or more extreme if the null is true
|
|
critical value
|
value of test statistic associated with alpha level (defines criticial region)
|
|
critical region
|
region that identifies values of the test statistic associated with statistically significant outcomes (if the observed value falls in critical region, the null is rejected; if not, the null is not rejected)
|
|
alpha level (significance level)
|
probability of obtaining an observed value in the critical region even if the null were true
|
|
statistically significant
|
the conclusion to reject the null hypothesis, made when the p-value is smaller than alpha
|
|
proportion of explained variability statistic
|
(in the DV) is indicated by the square of the effect size statistic
|
|
t-test
|
(for independent samples) population mean and SD is unknown
|
|
assumptions of a t-test
|
a. scores have been sampled randomly from the population, b. sampling distribution of the mean is normal, c. within-groups variances are homogeneous
|
|
example of t-test from book
|
Haney, 1984 two sample experiment on whether death-qualifying a jury affects how simulated jurors perceive a criminal defendant
|
|
z-test
|
inferential statistics with one sample compared to a population (with a known SD)
|
|
increasing sample size results in..
|
increase statistical significance of a relationship whenever the effect size is greater than zero
|
|
p-value is not a good indicator of a size of a relationship because..
|
it is strongly influenced by a sample size, so smaller p-values do not mean there is a stronger relationship, just means the results are even less likely due to chance
|
|
effect size is the best index of the strength of a relationship because..
|
it is not influenced by the sample size. it stays the same regardless of the sample size.
|
|
Pearson Correlation coefficient (r)
|
used to evaluate direction and strength of relationship b/w two variables (measured on an interval or ratio scale)-ranges from +1 to -1 and represents effect size statistic for correlation. p-value of r can be determined and t-test used to analysis of data.
|
|
linear relationship
|
a relationship b/w two variables that can be approximated with a straight line
|
|
nonlinear relationship
|
relationship b/w variables that cannot be approximated with a straight line
|
|
positive correlation
|
as one variable increases, the other variable increases; as one variable decreases, the other variable decreases
|
|
negative correlation
|
as one variable decreases, the other variable increases
|
|
no correlation
|
r=0
|
|
independent-no correlation
|
no relationship b/w the two variables (can't use one variable to predict the other)
|
|
curvilinear-no correlation
|
relationships b/w two variables that change direction, and thus are not described by a single straight line (ex. stress and performance)
|
|
coefficient of determination (r squared)
|
proportion of variance in one variable explained by the other variable (ex. an r of .6 would have a coefficient of determination of .36)
|
|
restriction of range
|
size of r most likely decreases if there is a restriction of range (occurs when most participants have similar scores on one of the variables being correlated)
|
|
linear regression
|
a statistical technique for analyzing a research design in which only one predictor variable is used to predict a single outcome variable
|
|
regression line
|
line that best represents or fits the data (can be described mathematically by equation for straight line)
|
|
regression line y variable
|
outcome variable
|
|
regression line x variable
|
predictor variable
|
|
null hypothesis in correlational research
|
r=0, no relationship between the two variables
|
|
alternative (research) hypothesis of correlational research
|
r doesn't equal zero; there is a relationshpi between the to variables
|
|
significant r
|
indicates linear relationship that can be used to predict one variable from the other
|
|
spearman correlation coefficient
|
used to measure relationships between two variables measured on ordinal scales. uses rankings to compute correlation coefficient.
|
|
chi-square statistic
|
used to assess the relationship between two nominal variables
|
|
chi square test of independence
|
tests the likelihood that two variables are independent (when the frequency distribution for one variable is NOT related to the categories of the second variable). (ex. can measure relationship b/w political affiliation and atttitudes toward nuclear power)
|
|
contingency table
|
displays the number of individuals in each of the combinations of the two variables
|
|
null hypothesis for chi-square
|
the frequency distribution for one variable will have the same shape (same proportion) for all categories of the second variable (ex. the proportion of democrats and republicans will be the same for all categories of attitudes toward nuclear power)
|
|
alternative (research) hypothesis for chi-square
|
the frequency distribution for one variable will have a different shape (different proportion) for all categories of the second variable (ex. the proportion of democrats and republicans will not be the same for all categories of attitudes toward nuclear power)
|
|
chi-square analysis
|
compare contingency table of observed values to expected values (if the null were true). if difference is big enough (statistically significant chi-square) then reject the null and conclude there is a relationship between the two variables)
|
|
multiple regression
|
a statistical technique for analyzing a research design in which more than one predictor variable is used to predict a single outcome variable (tested with an F-test)
|
|
zero-order correlations
|
correlations (r values) b/w each of the predictor variables and the outcome variable
|
|
beta weights (regression coefficients)
|
statistics that indicate the relationship between each of the predictor variables and the outcome variable holding constant the effects of the other predictors
|
|
multiple correlation coefficient R
|
(effect size statistic for multiple regression) - a statistic that indicates the extent to which all of the predictor variables in a regression analysis are able together to predict the outcome variable
|
|
R squared
|
represents the proportion of variance in the outcome variable explained by all of the predictor variables together
|
|
path analysis
|
(often used with longitudinal research designs) assesses the relationships among a number of variables and examines the possible causal directions among the set of correlated variables
|
|
path diagram
|
a graphic display of the relationships among a number of variables (results of the path analysis displayed visually)
|
|
Reverse causation
|
(directionality problem – does x cause y or does y cause x?) – the possibility that the outcome variable causes the predictor variable rather than the predictor variable causing the outcome variable
|
|
Reciprocal causation
|
the possibility that the predictor variable causes the outcome variable AND the outcome variable also causes the predictor variable
|
|
Third (common-causal) variable
|
variables that cause both the predictor variable and the outcome variable
|
|
Spurious relationship
|
a relationship between two variables that is produced by a common-causal (third) variable
|
|
Extraneous variables
|
variables other than the predictor variable that cause the outcome variable but that do NOT cause the predictor variable
|
|
Mediating variables (mediator)
|
a variable that is caused by one variable that in turn causes another variable
|
|
Correlation matrix
|
a table showing the correlations of many variables with each other
|
|
Panel studies (longitudinal research designs)
|
research in which the same individuals are measured more than one time and the time period between the measurements is long enough that changes in the variables of interest could occur
|
|
Cross-sectional research design
|
research in which comparisons are made across people from different age groups, but all groups are measured at the same time
|
|
Structural equation analysis
|
a multivariate statistical procedure that tests whether the actual relationships among a set of collected variables conform to a theoretical prediction about how those variables should be related
|
|
Latent variables
|
the conceptual variables or factors in a structural equation analysis
|
|
experimental condition(s)
|
the level of the independent variable in which the situation of interest was created (can have more than one experimental condition)
|
|
control condition(s)
|
the level of the independent variable in which the situation of interest was NOT created (can have more than one control condition; some experiments do NOT have a control condition at all [e.g., imagery vs. rote rehearsal])
|
|
Levels of an IV
|
the specific situations created by the experimental manipulation
|
|
conditions
|
a term used to describe the levels of an experimental manipulation in one-way experimental designs
|
|
Participant variables
|
a variable that represents differences among individuals on a demographic characteristic or a personality trait (e.g., ethnicity, gender)
|
|
Between-participants designs
|
experiments in which the comparison of the scores on the dependent variable is between the participants in the different levels of the independent variable and each individual is in only one level (each participant participates in only ONE level of the IV)
|
|
Within-participants (repeated-measures) designs
|
experiments in which the same people participate in more than one condition of an experiment, thereby creating equivalence, and the differences across the various levels are assessed within the same participants (each participant participates in ALL the levels of the IV)
|
|
random assignment
|
a method of ensuring that the participants in the different levels of the independent variable are equivalent before the experimental manipulation occurs
|
|
within-participants design
|
repeated measures
|
|
matched-groups design
|
research design in which participants are measured on a variable of interest before the experiment begins and are then assigned to conditions on the basis of their scores on that variable
|
|
Hypothesis testing in one-way experimental designs with Two-groups (two levels of IV)
|
t-test
|
|
t-test advantages and disadvantages
|
advantages – simple to conduct, analysis is easy; disadvantage – doesn’t tell you about function of relationship
|
|
Hypothesis testing in one-way experimental designs with Three or more groups (more than 2 levels of IV)
|
ANOVA
|
|
advantages and disadvantages for ANOVA
|
advantage – tells you about function of relationship (e.g., can identify curvilinear relationships between variables); disadvantages – more difficult to conduct, analysis is harder
|
|
Between-groups variance
|
a measure of the variability of the dependent variable across the experimental conditions in ANOVA (called mean-square between). treatment effect + individual differences + experimental error
|
|
F ratio – in ANOVA
|
a statistic that assesses the extent to which the means of the experimental conditions differ more than would be expected by chance
|
|
Within-groups variance
|
a measure of the variability of the dependent variable across the participants within the experimental conditions in ANOVA (called mean square within). individual differences + experimental error
C. F ratio – in ANOVA – a statistic that assesses the extent to which the means of the experimental conditions differ more than would be expected by chance |
|
Degrees of freedom
|
number of values that are free to vary given restrictions that have been placed on the data
|