Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
50 Cards in this Set
- Front
- Back
Descriptive Statistics
|
Numerical and pictorial information about variables.
Mean, mode, median, St.dev and range Statistics which describe the sample data without drawing inferences about the larger population. |
|
external validity
|
characteristics of a study that refers to the extent to which findings can be generalised to individuals, settings and conditions beyond the scope of the specific experiment
|
|
Co-variance
|
A measure of how much 2 variables change together
|
|
Correaltion
|
Shows the relationship between two variables. NOT causation!
|
|
Internal Validity
|
Changes to dependent variable are due to independent variable. Ensures an unambiguous interpretation for the outcome of the experiment.
|
|
Effect Size
|
A measure that indicates the strength of the relationship between the independent variable and the dependent variable
|
|
Meta- Analysis
|
Summary of results of more than one experiment on an important research problem
|
|
Type I error
|
Rejecting the null hypothesis when the null is true. (Saying something is meaningful when it is not)
|
|
Type II error
|
Accepting the null hypothesis when it is not true. (Saying something is not meaningful when it is)
|
|
Statistical inference
|
The process of estimates and conclusions carefully based on data from a sample.
|
|
Sample
|
A sub-group that is representative of the population
|
|
Population
|
The entire group of interest for a statistical conclusion
|
|
Extraneous variables
|
potential variables that are not of interest to the researcher.
|
|
Demand Characteristics
|
Participants interpret the experiment's purpose and unconsciously change their behaviour
|
|
Within Groups
|
Repeated measures: one group tested twice to see if there is a difference between conditions
Use dependent t-test |
|
When to use dependent t-test
|
Data is normally distributed
When there are 2 levels to the independent variable Within groups design |
|
Sig > .05
|
Probability that there is no difference
Stick with null hypothesis The assumption of normality/homogenity was not violated |
|
Sig < .05
|
Probability that there is a difference
Enough evidence to reject null hypothesis The assumption of normality/homogenity was violated |
|
Confidence interval depends on
|
sample size, sample mean and confidence level
|
|
A very small sig value allows...
|
the researcher to conclude that there is enough evidence to reject the null
|
|
Anova
|
Comparing the means. An expansion of t-test to more than 2 groups
|
|
Test of Homogenity
|
statistical test checking the assumption of equal SD in both groups
|
|
Post hoc test
|
Looking at data after experiment has concluded
Only use if there is a significant difference |
|
Causation
|
two variables are causally related if changes in the value of one cause the other to change.
|
|
Confidence Interval
|
An interval, with limits at either end, with a specified probability of including the parameter being estimated.
|
|
Point Estimate
|
Single value that represents the best estimate of the population value
|
|
Interval estimate
|
Builds on the point estimate to produce a range of values
|
|
Confirmation bias
|
A tendency to favour information that is consistent with pre-existing beliefs
|
|
Operational defintion
|
Defining a term for research accuracy
|
|
Alpha
|
the probability of Type I error
|
|
Boxplot
|
The graphical representation of the dispersion of a sampel
|
|
Categorical data
|
Data representing counts or number of observations in each category.
|
|
Chi-Square test
|
A statistical test often used for analysing categorical data.
|
|
Critical Value
|
The value of a test statistic at or beyond which we will reject H0 .
|
|
Inferential statistics
|
That branch of statistics that involves drawing inferences about parameters of the population(s) that have been sampled.
|
|
Interquartile range
|
The range of the middle 50% of the observations.
|
|
Kurtosis
|
A measure of the peakedness of a distribution.
|
|
Measures of central tendency
|
Numerical values referring to the centre of the distribution.
|
|
Mutually exclusive
|
Two events are mutually exclusive when the occurrence of one precludes the occurrence of the other.
|
|
Negative relationship
|
A relationship in which increases in one variable are associated with decreases in the other.
|
|
Negatively skewed
|
A distribution that trails off to the left.
|
|
One-Way Anova
|
An analysis of variance where the groups are defined on only one independent variable.
|
|
Outlier
|
An extreme point that stands out from the rest of the distribution.
|
|
Parametric tests
|
Statistical tests that involve assumptions about, or estimation of, population parameters.
|
|
Positively skewed
|
A distribution that trails off to the right.
|
|
Variance
|
The sum of the squared deviations from the mean, divided by the degrees of freedom (N- 1).
|
|
Dependent t-test
|
compares the means between 2 related groups on the same continuous variables
|
|
Assumptions of a dependent t-test
|
1. DV is measured on ratio/interval level
2. IV consists of 2 categorical related groups 3. No significant outliers in the difference between the two related groups 4. Distribution of the differences in the DV between the two groups in approx normally distributed |
|
One Way Anova Between Groups
|
stands for one way analysis of variance
used to determine whether there are any significant differences between the means of 2 or more independent groups |
|
Assumptions of One Way ANOVA
|
1. DV is measured on interval/ratio level
2. IV consists of 2 or more categorical independent groups 3. Independence of observation (no pp in more than one group) 4. No significant outliers 5. DV should be approx normally distributed for each category of the IV (Shapiro Wilk test) 6. Homeogenity of Variance (Levene's) |