• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/35

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

35 Cards in this Set

  • Front
  • Back

Alternate Forms Reliability

Method of estimating a test's reliability that entails administering two forms of the test to the same group of examinees and correlating the two sets of scores. Forms can be administered at about the same time (coefficient of equivalence) or at different times (coefficient of equivalence and stability). Some see this as the best method for assessing reliability. Difficulty in creating equivalent forms.
Classical Test Theory
Theory of measurement that regards observed variability in test scores as reflecting two components: true differences between the examinees on the attribute(s) measured by the test and the effects of measurement (random) error. Reliability is a measure of true score reliability.
Coefficient Alpha
Method for assessing internal consistency reliability that provides an index of average inter-item consistency.
Kuder-Richardson 20
Can be used as a substitute for coefficient alpha when test items are scored dichotomously.
Construct Validity
The extent to which a test measures the hypothetical trait (construct) it is intended to measure. Established through use of convergent and discriminant validity on other tests, factor analysis, determining if changes is scores reflect developmental changes, and seeing if experimental manipulations have the effect that is expected.
Content Validity
The extent to which a test adequately samples the domain of information, knowledge or skill that it purports to measure. Determined primarily by expert judgement. Important for achievement and job tests.
Criterion Contamination
Refers to bias introduced into a person's criterion score as a result of the knowledge of the scorer about performance on the predictor. Tends to artificially inflate the relationship between the predictor and criterion.
Criterion-referenced interpretation
Interpretation of a test score in terms of a prespecified standard; in terms of percent of content correct (percentage score) or of predicted performance on an external criterion (regression equation, expectancy table).
Criterion-related validity
The type of validity that involves determining the relationship (correlation) between the predictor and the criterion. The correlation coefficient is referred to as the criterion-related validity coefficient. Can be either concurrent or predictive.
Factor Analysis
A multivariate statistical technique used to determine how many factors (constructs) are needed to account for the intercorrelations among a set of tests, subtests, or test items. True score variability consists of communality and specificity. Factors can be orthogonal or oblique.
Factor Loading
In factor analysis, a factor loading is the correlation between a test ( or other variable included in the analysis) and a factor.
Communality
The communality is the total amount of variability in scores on the test (or other variable) that accounted for by the factor analysis.
Incremental Validity
The extent to which a predictor increases decision-making accuracy. Calculated by subtracting the base rate from the positive hit rate.
True Positives
Those who scored high on the predictor and the criterion.
False Positives
Those who scored high on the predictor and low on the criterion.
True Negatives
Scored low on the predictor and on the criterion.
False Negatives
Scored low on the predictor and high on the criterion.
Item Characteristics Curve
When using item response theory, an iten characteristic curve is constructed for each item by plotting the proportion of examinees who answered the item correctly against total test score, an external criterion, or a mathematically derived estimate of a latent ability or trait.
Item Difficulty
An item's difficulty level is calculated by dividing the number of individuals who answered the item correctly by the total number of individuals. Ranges from 0 to 1.0, .5 is preferred.
Item Discrimination
The extent to which an item differentiates between examinees who obtain a high score versus low scores. Ranges from -1.0 to 1.0.
kappa statistic
A correlation coefficient used to assess inter-rater reliability.
Multitrait-multimethod matrix
A systemic way to organize the correlation coefficients obtained when assessing a measure's convergent and discriminant validity. Measure at least two traits with at least two methods.
Norm-referenced interpretation
Interpretation of an examinee's test performance relatie to the performance of examinees in a normative (standardization) sample. Percentile ranks, standard scores, t-scores, and z-scores are norm referenced.
Orthogonal Rotation
In factor analysis, an orthogonal rotation of identified factors produces uncorrelated factors.
Oblique Rotation
In factor analysis, produces correlated factors.
Relationship between Reliability and Validity
Reliability is a necessary but not sufficient condition for validity.
Relevance
In test construction, relevance refers to the extent to which test items contribute to achieving the stated goals of testing.
Reliability
The consistency of test scores across time, across items, or over different forms. Methods include test-retest, alternative forms, split-half, coefficient alpha, and inter-rater.
Reliability Coefficient
Interpreted directly as a measure of true score variability. .80 means that 80% of the variability is due to true score variability.
Split-half Reliability
Assessing reliability by splitting the test in half and correlating the scores on two halves.
Spearman-Brown Formula
Correction to split-half coefficient. which estimates what the test's reliability would be if it were based on the full length of a test.
Standard Error of Estimate
An index of error when predicting criterion scores from predictor scores. Used to construct a confidence interval. Depends on the criterion's standard deviation and the predictor's validity coefficient.
Standard Error of Measurement
An index of measurement error. Used to construct a confidence interval around an examinees obtained test score. depends on test's standard deviation and reliability coefficient.
Increase Reliability
Increase test length, increase heterogeneity of sample, increase the range of obtained scores.
Test-retest Reliability
A method for assessing reliability that involves administering the same test to the same group of examinees on two different occasions and correlating the two sets of scores. Yields a coefficient of stability.