• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/53

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

53 Cards in this Set

  • Front
  • Back
NASP Principle of Ethical and Legal Consideration in Measurement
Principle II.3: Responsible Assessment and Intervention Practices
-Highest standard for responsible professional practice in assessment and intervention
APA Ethical Code for Measurement
Standard 9
-Bases for assessment, use of assessments, informed consent, release of test data, test construction, interpretation of assessment results, assessment by unqualified persons, obsolete tests, outdated test results, test scoring and interpretation services, explaining assessment results, maintaining test security
IDEA, 2004 in regards to Ethical and Legal Considerations in Measurement
-Funds to state and local educational agencies to provide free and appropriate public education (FAPE)
-Legal requirements for evaluating and identifying children with disabilities: Need for special education must be established and a disability in obtaining an education must be documented
What is nondiscriminatory assessment and classification?
Multifaceted assessment process, awareness of test bias, due process, equal protection
What is informed consent?
Written parental consent
What are some considerations needed in assessment planning?
Multifaceted, comprehensive, fair, valid, useful
What needs to be looked at when selecting instruments?
Reliability, validity, adequacy of norms
What are some aditional considerations needed in measurement?
-Nondiscriminatory assessment and classification
-Informed consent
-Assessment planning
-Selection of instruments
-Competence
-Data collection
-Data interpretation
-Making recommendations
-Report writing
-Sharing results
-IEP teams
What are the 5 stages in developing, norming, and revising tests?
1. Test Conceptualization: What is test designed to measure, objective, need, population, format, training, interpretation
2. Test Construction: Scales decided on based on what best suits the measurement of variable being tested
3. Test tryout: Tried on people similar in critical aspects to people for whom it is designed, identical conditions, eliminate extraneous variables
4. Test Analysis: Goal--select best items from pool of "tryout" items, consider index of items' difficulty, reliability, validity, discrimination
5. Test Revision: Throw out bad items, administer to appropriate sample under standardized conditions, norming is important
What is Standardization
A process to ensure uniformity in test administration, scoring, and interpretation
What are some reasons that tests may need revision?
-Stimulus materials look dated
-Verbal content not readily understood by test-takers
-Culture changes
-Reliability/validity improvements
-Theory on which test is based has been improved
What is reliability?
Consistency of test scores from one testing session to another
What are some sources of error variance?
-Test construction
-Test administration
-Test scoring and interpretation
What is test-retest reliability?
-Using same instrument to measure the same thing at two points in time
What are some sources of error in measuring test-retest reliability?
-Greater amount of time between--lower reliability coefficient
-Experience, practice effect, memory, fatique, motivaton
-
Explain parallel/alternate forms in determining reliability.
-Different, equivalent versions
-Minimizes effect of memory
-Measures coefficient of equivalence
Explain internal consistency/split-half reliability
-Correlating two pairs of scores from equivalent halves of same test
-Measure of internal validity and internal consistency
What is inter-scorer reliability?
-Degree of agreement between two scorers
-Indicates that it may be scored with consistency with proper training
What is Reliability of Individual Scores?
-Standard error of measurement (percision)
-Confidence interval (range/band likely to contain true score)
What are some methods used to determine reliability of scores?
-Test-retest
-Parallel/alternate forms
-Internal consistency/split-half
-Inter-Scorer
-Reliability of Individual Scores
What is Validity?
Degree to which an instrument measures what it purports to measure
What is content validity?
-Degree to which a tool samples behavior representative of the behavior it is designed to measure
-How adequately questions represent content being assessed
-Accuracy of test format
What is criterion-related validity?
-Degree to which how adequately a score on a test can be used ot infer an individual's standing on some criterion
-Compare performance on one test with performance on another
-Degree of relationship between scores--validity coefficient
What is concurrent criterion validity?
-Degree to which test score is related to some criterion measure obtained at same time
What is predictive criterion validity?
Degree to which test score predicts performance on criterion measure obtained in the future or on expectancy table
What is construct validity?
-Broadest category
-Judgment about appropriateness of inferences drawn from test scores regarding individual standings on a variable called a construct (scientific idea developed to describe or explain behavior)
-Measured by assessing evidence on predictions made from theory
What are two different Types of Tests/Measures?
Norm-Referenced Tests & Criterion-Referenced Tests
What are norm-referenced tests?
Compare test taker's score to scores of a group of test takers
What are criterion-referenced tests?
Assess whether the test taker has met certain criteria
What is the use of norm-referenced tests?
Survey individual differences
What are the use of criterion-referenced tests?
Mastery of task, competence
How are norm-referenced tests interpreted?
Compare individuals
How are criterion-referenced tests interpreted?
Compare performance to criterion
What is the content of norm-referenced tests?
Broad
What is the content of criterion-referenced tests?
Narrow
How is performance measured in norm-referenced tests?
Relative position
How is performance measured in criterion-referenced tests?
Absolute
List some intelligence tests
-Stanford-Binet 5th Ed. (SB-5)
-Wechsler (WISC-IV), WPPSI III, WAIS IV)
-Kaufman Assessment Battery for Children (KABC)
-Woodcock-Johnson III (WJ III-COG)
-Differential Ability Scales (DAS)
List some academic achievement tests
-Kaufman Test of Educational Achievement (KTEA)
-Woodcock-Johnson III (WJ III-ACH)
-Woodcock Reading Mastery Tests-Revised (WRMT-R)
-Gray Oral Reading Test (GORT)
-KeyMath
List some Personality/Projective Tests
-Self-Direct Search (SDS)
-Minnesota Multiphasic Personality Inventory (MMPI)
-Rorschach Inkblots
-Thematic Apperception Test (TAT)
List some Clinical/Counseling Tests
-Beck Depression Inventory
-Parenting Stress Index (PSI)
-Structured Interviews
List some test for individuals with Behavior Disabilities
-Connors III
-Achenbach Child Behavior Checklist (CBCL)
-Vineland
-Behavior Assessment Scale for Children (BASC-2)
List some Career/Business Tests
-Checklist for Adaptive Living Skills (CALS)
-Myers-Briggs Type Indicator (MBTI)
List the pros of the discrepancy test model
-Objectivity
-Easy to apply and understand
-Statistical properties to establish predictable LD population size
-Allows examiners to evaluate learning style and information processing skills during testing
List the cons of the discrepancy test model
-Not practical or accurate for young students
-Requires a student to each a certain level of failure before being identified
-Does not adequately consider error of testing--will identify non-LD students while failing to identify others
-Time-consuming process
List the pros of the RTI Model
-Can be applied to students of all ages
-Does not require student to fail for a significant period of time before being identified
-Forces schools to be proactive in providing mainstream interventions
-Requires little educational disruption for testing
List the cons of the RTI Model
NONE
What is a CBM and who identified this definition?
-A set of methods for indexing competence and growth on basic academic skills
-Elliot and Fuchs (1997)
What does a CBM entail?
-Sampling test stimuli from local curricula and administering and scoring on a routine basis
-Produce reliable and valid information about students' academic skills for English-only and ELL students
-When administered routinely, alternate forms of CBM models academic growth well and enhances instructional planning.
What are the benefits of using a CBM?
-Longitudinal measurement
-Informing instructional planning
-Documenting treatment effects
Define Longitudinal Measurement
1. Taps a construct that is constant in both quality and difficulty
2. Summarizes performance on same raw score metric and provides equal scaling over time
3. Provides enough alternate-forms for the whole year
Explain how CBMs inform instructional planning
-Wesson (1991) and Fuchs, Deno, and Mirkin (1984) provide evidence that student outcomes improve when special educators use CBM to inform instructional planning
Explain how CBMs can be used to document treatment effects
-Marston, Fuchs & Deno (1986) documented that CBM reigstered more student growth than traditional in 10-week period
-Fuchs, Fuchs, and Hamlett (1989) demonstrate CBM slopes reflected treatment effects more sensitively than traditional pre-post measurement