• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/29

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

29 Cards in this Set

  • Front
  • Back

measurement

the assignment of #s to represent the amt of an attribute present in an object/person, using specific rules

advantages to measurement

removes guesswork, provides precise info, less vague than words (#s allow you to do statistical analyses & get answers)

Nominal measurement

assigning #s to classify characteristics into categories; interchangeable


lowest level of evidence b/c it only uses #s to categorize so they have no meaning

Ordinal Measurement

ranking objects based on their relative standing on an attribute; you can't make inferences; quantitative ordering of values


e.g., educational level

interval measurement

objects ordered on a scale that has equal distances b/n points on the scale; # has meaning


e.g., depression scale, likert scale

Ratio measurement

equal distances b/n score units; there is a rational, meaningful zero; highest level of evidence; gives you actual #s


e.g., BP, wt, age, etc.

obtained score

an actual data value for a participant


e.g., anxiety scale score


obtained score = true score +/- error

True score

the score that would be obtained with an infallible measure

Error

the error of measurement, caused by factors that distort the measurement

4 factors contributing to errors of measurement

1. situational contaminants (temp/smell of environment that affects how honestly & purposefully subjects answer)


2. transitory personal factors (e.g., fatigue)


3. response set biases (social desirability, nay-sayers, extreme responses


4. item sampling (what was used to measure the variable?)

reliability

the consistency & accuracy w/which an instrument measures the target attribute; being able to use the same instrument multiple times and getting the same answer

reliability coefficient

used to compute reliability assessments; done by a computer


-range from .00-1.00 (1 is perfect score; never seen)


<.70 considered unsatisfactory


.80+ are desirable

What 3 aspects of reliability do we evaluate?

stability, internal consistency, equivalence

stability

the extent to which scores are similar on 2 separate administrations of an instrument




-evaluated by test-retest reliability


-requires participant to complete the same instrument on 2 occasions


-appropriate for relatively enduring attributes (e.g., creativity)

Internal consistency

the extent to which all the items on an instrument are measuring the same attribute


-evaluated by administering instrument on one occasion


-appropriate for most multi-item instruments


*most widely used approach to assess reliability


-assessed by computing coefficient alpha (Cronbach's alpha)


Alphas >/= .80 highly desirable

Equivalence

the degree of similarity b/n alternative forms of an instrument or b/n multiple raters/observers using an instrument




-most relevant for structured observations


-assessed by comparing agreement b/n observations or ratings of 2+ observers (interobserver/interrater reliability)

Reliability Principles

-low reliability can undermine adequate testing -of hypotheses


-reliability estimates vary depending on procedure used to obtain them


-lower in homogeneous vs. heterogeneous samples


-lower in shorter vs. longer multi-item scales

Validity

the degree to which an instrument measures what it's supposed to measure


-aspects: face validity, content validity, criterion-related validity, construct validity

face validity

refers to whether the instrument looks as though it is an appropriate measure of the construct


-based on judgment; no objective criteria for assessment

content validity

the degree to which an instrument has an adequate sample of items for the construct being measured


-evaluated by expert evaluation, often via a quantitative measure- the content validity index (CVI)


0.9+ is desirable

criterion-related validity

the degree to which the instrument is related to an external criterion


-validity coefficient: calculated by analyzing the relationship b/n scores on the instrument and the criterion


-predictive and concurrent validity



predictive validity

the instrument's ability to distinguish ppl whose performance differs on a future criterion (e.g,. Kaplan test predicting NCLEX scores)



concurrent validity

the instrument's ability to distinguish individuals who differ on a present criterion; substituting one measure for another (e.g., AP class exams earning college credit)


-one score replaces another

construct validity

degree to which the measure you're using is addressing the theoretical construct you want to measure


-what is the instrument really measuring?


-Does it adequately measure the construct of interest?


-Convergent & discriminant validity (are they converging in what they measure of are they 2 discriminant things?)

methods of assessing construct validity

known-groups technique, testing relationships based on theoretical predictions, factor analysis

known-groups technique

2 or more groups that know they should be diff when given the same instrument; test is invalid is the answer is the same

factor analysis

complex process done by computer analysis; testing an attribute to place like terms into subgroups

sensitivity

the instrument's ability to correctly identify a "case" (e.g., diagnose a condition)


-e.g., how sensitive is this test at diagnosing strep throat? Am I getting a true positive?

Specificity

the instrument's ability to correctly identify non-cases, to screen out those w/o the condition


e.g., tests the true negative