Study your flashcards anywhere!
Download the official Cram app for free >
 Shuffle
Toggle OnToggle Off
 Alphabetize
Toggle OnToggle Off
 Front First
Toggle OnToggle Off
 Both Sides
Toggle OnToggle Off
 Read
Toggle OnToggle Off
How to study your flashcards.
Right/Left arrow keys: Navigate between flashcards.right arrow keyleft arrow key
Up/Down arrow keys: Flip the card between the front and back.down keyup key
H key: Show hint (3rd side).h key
A key: Read text to speech.a key
113 Cards in this Set
 Front
 Back
sampling distribution

distribution of all possible values of a statistic (a picture of the null hypothesis in hypothesis testing)


statistic

a value summarizing a measurable characteristic of a sample (mean of sample, SD of sample, etc.)


parameter

a number that represents the characteristics of a population (mean of population, SD of population, etc.)


sampling distribution of the mean

distribution of all possible means of samples of a given size taken from a population (tends to be normal in shape no matter what the shape is of the original distribution, especially for n>30)


binomial distribution

sampling distribution for events that have two equally likely possibilities (ex: coin flip)


sampling error

amount of error that exists between a sample statistic and a population parameter


standard error of the mean

standard deviation of the means in the sampling distribution of the mean (estimate of the amount of variability in the expected sample means in the sampline distribution of means)


null hypothesis (Ho)

the assumption that observed data reflect only what would be expected from the sampling distribution (no effect of treatment, no relationship b/w variables, no difference b/w means, etc.)


research (alternative) hypothesis (H1)

a prediction regarding the relationship b/w or among two or more variables


observed value

value of inferential test statistic based on observed data


pvalue (probablility value)

probability of obtaining an observed value (test statistic) as extreme or more extreme if the null is true


critical value

value of test statistic associated with alpha level (defines criticial region)


critical region

region that identifies values of the test statistic associated with statistically significant outcomes (if the observed value falls in critical region, the null is rejected; if not, the null is not rejected)


alpha level (significance level)

probability of obtaining an observed value in the critical region even if the null were true


statistically significant

the conclusion to reject the null hypothesis, made when the pvalue is smaller than alpha


statistically nonsignificant

the conclusion to not reject the null hypothesis made when the pvalue is great than alpha


effect size

a statistic that indexes the size of a relationship


onesided (onetailed) pvalues

pvalues that consider only the likelihood that a relationship occurs in the predicted direction


twosided (twotailed) pvalues

p values that consider the likelihood that a relationship can occur either in the expected or in the unexpected direction (more commonly used)


steps in hypothesis testing

1. develop research hypothesis, 2. set alpha (usually a=.05), 3. collect data, 4. calculate test statistic (observed value) and corresponding pvalue, 5. compare pvalue to alpha


p < a (p<.05)

reject null


p > a (p>.05)

fail to reject null


type I error

("false alarm") occurs when we reject a true null (conclude a coil is unfair when it is fair) probability=a


type II error

("miss") occurs when we fail to reject a false null (conclude a coin is fair wen it is unfair) probability=b


correct decision I

occurs when we fail to reject a true null (conclude a coin is fair when it is fair) considered a "boring" result. probability=1a


correct decision II

("hit") occurs when we reject a false null (conclude a coin is unfair when it is unfair) *known as POWER (the ability of a test to fnd a difference when there really is one) probability=1b


relationship between a and b

when a is lower (probability of making a type I error is lower) then b is higher (probability of making a type II error is higher)


trade off between type I and type II errors

most researchers believe it is worse to make a type I error than a type II error so alpha is ALMOST ALWAYS set lower than beta


ttest (for independent samples)

population mean and SD is unknown


statistical significance

effect size x sample size


sampling distribution

distribution of all possible values of a statistic (a picture of the null hypothesis in hypothesis testing)


statistic

a value summarizing a measurable characteristic of a sample (mean of sample, SD of sample, etc.)


parameter

a number that represents the characteristics of a population (mean of population, SD of population, etc.)


sampling distribution of the mean

distribution of all possible means of samples of a given size taken from a population (tends to be normal in shape no matter what the shape is of the original distribution, especially for n>30)


binomial distribution

sampling distribution for events that have two equally likely possibilities (ex: coin flip)


sampling error

amount of error that exists between a sample statistic and a population parameter


standard error of the mean

standard deviation of the means in the sampling distribution of the mean (estimate of the amount of variability in the expected sample means in the sampline distribution of means)


null hypothesis (Ho)

the assumption that observed data reflect only what would be expected from the sampling distribution (no effect of treatment, no relationship b/w variables, no difference b/w means, etc.)


research (alternative) hypothesis (H1)

a prediction regarding the relationship b/w or among two or more variables


observed value

value of inferential test statistic based on observed data


pvalue (probablility value)

probability of obtaining an observed value (test statistic) as extreme or more extreme if the null is true


critical value

value of test statistic associated with alpha level (defines criticial region)


critical region

region that identifies values of the test statistic associated with statistically significant outcomes (if the observed value falls in critical region, the null is rejected; if not, the null is not rejected)


alpha level (significance level)

probability of obtaining an observed value in the critical region even if the null were true


statistically significant

the conclusion to reject the null hypothesis, made when the pvalue is smaller than alpha


proportion of explained variability statistic

(in the DV) is indicated by the square of the effect size statistic


ttest

(for independent samples) population mean and SD is unknown


assumptions of a ttest

a. scores have been sampled randomly from the population, b. sampling distribution of the mean is normal, c. withingroups variances are homogeneous


example of ttest from book

Haney, 1984 two sample experiment on whether deathqualifying a jury affects how simulated jurors perceive a criminal defendant


ztest

inferential statistics with one sample compared to a population (with a known SD)


increasing sample size results in..

increase statistical significance of a relationship whenever the effect size is greater than zero


pvalue is not a good indicator of a size of a relationship because..

it is strongly influenced by a sample size, so smaller pvalues do not mean there is a stronger relationship, just means the results are even less likely due to chance


effect size is the best index of the strength of a relationship because..

it is not influenced by the sample size. it stays the same regardless of the sample size.


Pearson Correlation coefficient (r)

used to evaluate direction and strength of relationship b/w two variables (measured on an interval or ratio scale)ranges from +1 to 1 and represents effect size statistic for correlation. pvalue of r can be determined and ttest used to analysis of data.


linear relationship

a relationship b/w two variables that can be approximated with a straight line


nonlinear relationship

relationship b/w variables that cannot be approximated with a straight line


positive correlation

as one variable increases, the other variable increases; as one variable decreases, the other variable decreases


negative correlation

as one variable decreases, the other variable increases


no correlation

r=0


independentno correlation

no relationship b/w the two variables (can't use one variable to predict the other)


curvilinearno correlation

relationships b/w two variables that change direction, and thus are not described by a single straight line (ex. stress and performance)


coefficient of determination (r squared)

proportion of variance in one variable explained by the other variable (ex. an r of .6 would have a coefficient of determination of .36)


restriction of range

size of r most likely decreases if there is a restriction of range (occurs when most participants have similar scores on one of the variables being correlated)


linear regression

a statistical technique for analyzing a research design in which only one predictor variable is used to predict a single outcome variable


regression line

line that best represents or fits the data (can be described mathematically by equation for straight line)


regression line y variable

outcome variable


regression line x variable

predictor variable


null hypothesis in correlational research

r=0, no relationship between the two variables


alternative (research) hypothesis of correlational research

r doesn't equal zero; there is a relationshpi between the to variables


significant r

indicates linear relationship that can be used to predict one variable from the other


spearman correlation coefficient

used to measure relationships between two variables measured on ordinal scales. uses rankings to compute correlation coefficient.


chisquare statistic

used to assess the relationship between two nominal variables


chi square test of independence

tests the likelihood that two variables are independent (when the frequency distribution for one variable is NOT related to the categories of the second variable). (ex. can measure relationship b/w political affiliation and atttitudes toward nuclear power)


contingency table

displays the number of individuals in each of the combinations of the two variables


null hypothesis for chisquare

the frequency distribution for one variable will have the same shape (same proportion) for all categories of the second variable (ex. the proportion of democrats and republicans will be the same for all categories of attitudes toward nuclear power)


alternative (research) hypothesis for chisquare

the frequency distribution for one variable will have a different shape (different proportion) for all categories of the second variable (ex. the proportion of democrats and republicans will not be the same for all categories of attitudes toward nuclear power)


chisquare analysis

compare contingency table of observed values to expected values (if the null were true). if difference is big enough (statistically significant chisquare) then reject the null and conclude there is a relationship between the two variables)


multiple regression

a statistical technique for analyzing a research design in which more than one predictor variable is used to predict a single outcome variable (tested with an Ftest)


zeroorder correlations

correlations (r values) b/w each of the predictor variables and the outcome variable


beta weights (regression coefficients)

statistics that indicate the relationship between each of the predictor variables and the outcome variable holding constant the effects of the other predictors


multiple correlation coefficient R

(effect size statistic for multiple regression)  a statistic that indicates the extent to which all of the predictor variables in a regression analysis are able together to predict the outcome variable


R squared

represents the proportion of variance in the outcome variable explained by all of the predictor variables together


path analysis

(often used with longitudinal research designs) assesses the relationships among a number of variables and examines the possible causal directions among the set of correlated variables


path diagram

a graphic display of the relationships among a number of variables (results of the path analysis displayed visually)


Reverse causation

(directionality problem – does x cause y or does y cause x?) – the possibility that the outcome variable causes the predictor variable rather than the predictor variable causing the outcome variable


Reciprocal causation

the possibility that the predictor variable causes the outcome variable AND the outcome variable also causes the predictor variable


Third (commoncausal) variable

variables that cause both the predictor variable and the outcome variable


Spurious relationship

a relationship between two variables that is produced by a commoncausal (third) variable


Extraneous variables

variables other than the predictor variable that cause the outcome variable but that do NOT cause the predictor variable


Mediating variables (mediator)

a variable that is caused by one variable that in turn causes another variable


Correlation matrix

a table showing the correlations of many variables with each other


Panel studies (longitudinal research designs)

research in which the same individuals are measured more than one time and the time period between the measurements is long enough that changes in the variables of interest could occur


Crosssectional research design

research in which comparisons are made across people from different age groups, but all groups are measured at the same time


Structural equation analysis

a multivariate statistical procedure that tests whether the actual relationships among a set of collected variables conform to a theoretical prediction about how those variables should be related


Latent variables

the conceptual variables or factors in a structural equation analysis


experimental condition(s)

the level of the independent variable in which the situation of interest was created (can have more than one experimental condition)


control condition(s)

the level of the independent variable in which the situation of interest was NOT created (can have more than one control condition; some experiments do NOT have a control condition at all [e.g., imagery vs. rote rehearsal])


Levels of an IV

the specific situations created by the experimental manipulation


conditions

a term used to describe the levels of an experimental manipulation in oneway experimental designs


Participant variables

a variable that represents differences among individuals on a demographic characteristic or a personality trait (e.g., ethnicity, gender)


Betweenparticipants designs

experiments in which the comparison of the scores on the dependent variable is between the participants in the different levels of the independent variable and each individual is in only one level (each participant participates in only ONE level of the IV)


Withinparticipants (repeatedmeasures) designs

experiments in which the same people participate in more than one condition of an experiment, thereby creating equivalence, and the differences across the various levels are assessed within the same participants (each participant participates in ALL the levels of the IV)


random assignment

a method of ensuring that the participants in the different levels of the independent variable are equivalent before the experimental manipulation occurs


withinparticipants design

repeated measures


matchedgroups design

research design in which participants are measured on a variable of interest before the experiment begins and are then assigned to conditions on the basis of their scores on that variable


Hypothesis testing in oneway experimental designs with Twogroups (two levels of IV)

ttest


ttest advantages and disadvantages

advantages – simple to conduct, analysis is easy; disadvantage – doesn’t tell you about function of relationship


Hypothesis testing in oneway experimental designs with Three or more groups (more than 2 levels of IV)

ANOVA


advantages and disadvantages for ANOVA

advantage – tells you about function of relationship (e.g., can identify curvilinear relationships between variables); disadvantages – more difficult to conduct, analysis is harder


Betweengroups variance

a measure of the variability of the dependent variable across the experimental conditions in ANOVA (called meansquare between). treatment effect + individual differences + experimental error


F ratio – in ANOVA

a statistic that assesses the extent to which the means of the experimental conditions differ more than would be expected by chance


Withingroups variance

a measure of the variability of the dependent variable across the participants within the experimental conditions in ANOVA (called mean square within). individual differences + experimental error
C. F ratio – in ANOVA – a statistic that assesses the extent to which the means of the experimental conditions differ more than would be expected by chance 

Degrees of freedom

number of values that are free to vary given restrictions that have been placed on the data
