Use LEFT and RIGHT arrow keys to navigate between flashcards;
Use UP and DOWN arrow keys to flip the card;
H to show hint;
A reads text to speech;
21 Cards in this Set
- Front
- Back
Alternate forms reliability
|
reliability obtained by administering two equivalent tests to the same group of examinees
|
|
Concurrent validity
|
Refers to how precisely a person's present performance (e.g. a test score) estimates that person's performance on the criterion measure at approximately the same time **correlates with current performance
|
|
Construct validity
|
the extent to which a test measures a theoretical construct or attribute ** i.e. intelligence tests- do they measure intelligence? intellect? achievement? what about cultural differences? judgement calls? intelligence is the theoretical construct for these tests.
|
|
Constructs
|
theoretical concepts, such as self-esteem and intelligence that can be obsered by some type of instrument
|
|
Content validity
|
refers to whether the individual items of a test represent what you actually want to assess **i.e. does a test question relate directly to a standard
|
|
Convergent validity
|
a test that has good convergent validity has high positive correlations with other tests measuring the same construct **i.e. an academic achievement test should correlate highly with established academic tests rather than with social and cognitive measures
|
|
Criterion-related validity
|
a method for assessing the validity of an instrument by comparing its scores with another criterion known already to be a measure of the same trait or skill **the extent to which a test measures or can predict some type of criteria (can be concurrent or predictive)
|
|
Discriminant validity
|
a test that has good discriminant validity has low correlations with tests that maeasure different constructs **i.e. an academic achievement test should correlate highly with established academic tests rather than with social and cognitive measures (academic test has discriminant validity with social and cognitive measures)
|
|
Interrater reliability
|
Involves having two raters independently observe and record specified behaviors (i.e. two teachers rating written portions of NYS tests)
|
|
Obtained score
|
the score actually calculated in the assessment process
|
|
Predictive validity
|
the extent to which a procedure allows accurate predictions about a subject's future behavior. It is a measure of a specific instrument's ability to predict future performance on some other measure or criterion at a later date **i.e. SAT or GRE exams to predict success in college or graduate school
|
|
Reliability
|
refers to the consistency of measurements
|
|
Reliability coefficient
|
expresses the degree of consistency in the measurement of test scores. ranges in value from 0 to 1. acceptable reliability coefficient should never be below 0.9
|
|
Reliable test scores
|
a test score that produces similar scores a cross various conditions and situations, including different evaluators and testing environments
|
|
Split-half reliability
|
indicates that subjects' scores on some trials consistently match their scores on other trials **also could be similar scores on 2 halves of the same test
|
|
Target behavior
|
a specific behavior an observer is looking to record
|
|
test-retest reliability
|
suggests that subjects tend to obtain the same score when tested at different times **f test is given multiple times, scores will be roughly the same
|
|
validity
|
the extent to which a test measures what it is supposed to measure
|
|
validity coefficient
|
criterion-related validity is usually expressed as a correlation between the test in question and the criterion measure. this correlation coefficient is referred to as a validity coefficient
|
|
Factors affecting validity
|
1. test related- anxiety, motivation, speed, understanding of test instructions, rapport, physical handicaps, language barriers, deficiencies in educational opportunities, and unfamiliarity with testing materials
2. establishment of criterion- if your comparison instrument is not valid, then the results may not be valid 3. intervening events- life experiences i.e. death of parent, divorce, breakup, move to new school, etc 4. reliability- if a test has low reliability then it will have low validity as well |
|
Factors affecting reliability
|
1. test length- the more homogeneous items there are, the greater the reliability
2. test-retest interval- smaller time interval between giving the test, the smaller the chance of change, and therefore the higher the reliability 3. variability of scores- the greater the variance of scores on a test, the higher the reliability is likely to be. small changes in performance have a greater impact on the reliability when the range of scores is narrow 4. guessing- the less guessing that occurs, the higher the reliability 5. variation within the test situation- the fewer variations in the way the test is administered, the higher the reliability (includes misleading or misunderstanding of directions, scoring errors, illness and daydreaming) |