• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/21

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

21 Cards in this Set

  • Front
  • Back
Alternate forms reliability
reliability obtained by administering two equivalent tests to the same group of examinees
Concurrent validity
Refers to how precisely a person's present performance (e.g. a test score) estimates that person's performance on the criterion measure at approximately the same time **correlates with current performance
Construct validity
the extent to which a test measures a theoretical construct or attribute ** i.e. intelligence tests- do they measure intelligence? intellect? achievement? what about cultural differences? judgement calls? intelligence is the theoretical construct for these tests.
Constructs
theoretical concepts, such as self-esteem and intelligence that can be obsered by some type of instrument
Content validity
refers to whether the individual items of a test represent what you actually want to assess **i.e. does a test question relate directly to a standard
Convergent validity
a test that has good convergent validity has high positive correlations with other tests measuring the same construct **i.e. an academic achievement test should correlate highly with established academic tests rather than with social and cognitive measures
Criterion-related validity
a method for assessing the validity of an instrument by comparing its scores with another criterion known already to be a measure of the same trait or skill **the extent to which a test measures or can predict some type of criteria (can be concurrent or predictive)
Discriminant validity
a test that has good discriminant validity has low correlations with tests that maeasure different constructs **i.e. an academic achievement test should correlate highly with established academic tests rather than with social and cognitive measures (academic test has discriminant validity with social and cognitive measures)
Interrater reliability
Involves having two raters independently observe and record specified behaviors (i.e. two teachers rating written portions of NYS tests)
Obtained score
the score actually calculated in the assessment process
Predictive validity
the extent to which a procedure allows accurate predictions about a subject's future behavior. It is a measure of a specific instrument's ability to predict future performance on some other measure or criterion at a later date **i.e. SAT or GRE exams to predict success in college or graduate school
Reliability
refers to the consistency of measurements
Reliability coefficient
expresses the degree of consistency in the measurement of test scores. ranges in value from 0 to 1. acceptable reliability coefficient should never be below 0.9
Reliable test scores
a test score that produces similar scores a cross various conditions and situations, including different evaluators and testing environments
Split-half reliability
indicates that subjects' scores on some trials consistently match their scores on other trials **also could be similar scores on 2 halves of the same test
Target behavior
a specific behavior an observer is looking to record
test-retest reliability
suggests that subjects tend to obtain the same score when tested at different times **f test is given multiple times, scores will be roughly the same
validity
the extent to which a test measures what it is supposed to measure
validity coefficient
criterion-related validity is usually expressed as a correlation between the test in question and the criterion measure. this correlation coefficient is referred to as a validity coefficient
Factors affecting validity
1. test related- anxiety, motivation, speed, understanding of test instructions, rapport, physical handicaps, language barriers, deficiencies in educational opportunities, and unfamiliarity with testing materials
2. establishment of criterion- if your comparison instrument is not valid, then the results may not be valid
3. intervening events- life experiences i.e. death of parent, divorce, breakup, move to new school, etc
4. reliability- if a test has low reliability then it will have low validity as well
Factors affecting reliability
1. test length- the more homogeneous items there are, the greater the reliability
2. test-retest interval- smaller time interval between giving the test, the smaller the chance of change, and therefore the higher the reliability
3. variability of scores- the greater the variance of scores on a test, the higher the reliability is likely to be. small changes in performance have a greater impact on the reliability when the range of scores is narrow
4. guessing- the less guessing that occurs, the higher the reliability
5. variation within the test situation- the fewer variations in the way the test is administered, the higher the reliability (includes misleading or misunderstanding of directions, scoring errors, illness and daydreaming)