• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/79

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

79 Cards in this Set

  • Front
  • Back
1. Define and be able to apply the broad definition of validity.
a. The degree to which a test actually measures what it purports to measure
2. What are the three main types of validity evidence?
Construct-related
• Does the test measure what it purports to measure?
– Criterion-related
• Does the test correlate with/predict one’s score on an external criterion?
– Content-related
• How adequately do test items measure the construct domain?
3. What prerequisites exist for validity?
a. Reliability is NECESSARY (but not sufficient) for validity
Validation studies are necessary to justify inferences that the test is valid (measures what it purports to ` measure).
4. Define Face Validity. How does it differ from other aspects of validity?
Mere Appearance that a measure measures what it is designed to measure; it differs because it does not offer evidence to support conclusions drawn from test scores
5. Define Content Validity: How is it measured?
Degree to which a test or measure adequately represents the conceptual domain (universe) it is designed to measure
Measured by the CVR and uses a panel of judges on a 3-point scale.
6. How do construct underrepresentation and construct-irrelevant variance relate to content validity?
Underrepresentation: Failure to capture important components of a construct; Construct Irrelevant variance: Scores are influenced by factors irrelevant to the construct.
7. What is the content validity ratio and how is it calculated?
a. uses a panel of judges to rate items on a 3-point
b. scale (Essential, Useful but not essential, Not necessary)
c. – >50% of panelists consider an item “essential”, it has some content validity
d. – The more judges who rate an item “essential” = higher its content validity
e. – The fewer judges who rate an item as “essential” = lower its content validity

CVR= (#of panelist rating item as essential- (N:total number of panelist/2))/ (N/2)
8. What is a criterion? What is Criterion-related Validity?
the standard against which a test is compared; How well test scores correspond with a particular criterion
9. Name and define the three subtypes of criterion related validity. Be able to give examples.
Predictive validity: accuracy with which test scores predict a criterion OBTAINED AT A LATER TIME
Ex: SAT is predictor; college GPA is criterion
Concurrent validity: Degree to which the test scores are related to the criterion and can be measured AT THE SAME TIME (Job samples)
Postdictive validity: accuracy with which a test score predicts a previously obtained criterion (test of antisocial behavior to predict past delinquent acts)
10. What is the validity coefficient? What is the meaning of a squared validity coefficient?
a. Relationship between a test and a criterion: Extent the test is valid for making statements about the criterion
b. Percentage of variation in the criterion we can expect to know in advance because of our knowledge of test scores.
11. What is incremental validity?
a. If test has unique predictive power
12. What is a construct? What is Construct-related Validity?
Construct: Something built by mental synthesis
Construct related Validity: Degree to which a set of measurement operations measures hypothesized constructs.
13. What are the two types of evidence in construct validity?
Convergent and discriminate
Convergent is a demonstration of similarity: measures of same construct converge or narrow in on the same thing. Ex: show a test measures same things as the other tests do for the same purpose. Or Demonstrate specific relationships one can expect if the test is doing its job.
Discriminant evidence: Demonstration of uniqueness: Low correlations with measures of unrelated constructs or evidence for what the test does NOT measure
Measure doesn’t represent a construct other than its devised purpose
14. How the two types of evidence for construct-validity defined and how are they different from one another?
a. Convergent and discriminate
i. Convergent is a demonstration of similarity: measures of same construct converge or narrow in on the same thing. Ex: show a test measures same things as the other tests do for the same purpose. Or Demonstrate specific relationships one can expect if the test is doing its job.
b. Discriminant evidence: Demonstration of uniqueness: Low correlations with measures of unrelated constructs or evidence for what the test does NOT measure
Measure doesn’t represent a construct other than its devised purpose
15. What is the relationship between reliability and validity?
Reliability is necessary but not sufficient in validity
16. Which two types of validity are logistical and not statistical? Why?
Face validity and content validity are logistical. They require good logic, intuitive skills and perseverance rather than statistics.
17. Which type of validity has been referred to as “the mother of all validities”, or “the big daddy” and why?
a. Construct related Validity: subsumes all activities used in the other types of validity; all validation is one and in a sense, all is construct validation.
19. What role does the relationship between examiner and test taker play?
a. Test scores are affected by examiner behavior and relationship.
20. What is the relationship between test examiner race and intelligence scores?
a. Race doesn’t really affect IQ but one can still communicate a hostile/friendly atmosphere…etc
21. Why would examiner race effects be smaller on IQ tests than on other psychological tests?
There is high structure to IQ tests; test procedures are very specific and not real any places to deviate.
22. What is the standard for test takers who are fluent in two languages?
Give test in their BEST language
23. Define expectancy effects and know whose name is associated with these effects
Data is affected by what an experimenter EXPECTS to find. Rosenthal is associated with effect.
24. A review of many studies showed that expectancy effects exist in ________ situations
Some, BUT NOT ALL
25. What types of situations might require the examiner to deviate from standardized testing procedures
When the person is blind/deaf or have other sorts of misunderstandings
26. What advantages and disadvantages were mentioned in lecture and the text regarding computer administered tests?
a. ADVANTAGES
i. NO JUDGEMENT PROBLEM
ii. Excellence of standardization.
iii. – Individually tailored sequential administration.
iv. – Precision of timing responses.
v. – Release of human testers for other duties.
vi. – Test taker not rushed.
vii. – Control of bias.
viii. – Objective and cost effective.
b. Disadvantage
i. Computer generated reports
ii. Hard to detect error if validation is inadequate
iii. People let the computer do too much of the thinking
iv. Lack of personal interaction
27. What subject variables can impact testing?
Test Anxiety
Worry; Emotionality; Lack of self-confidence
Illness; Elderly; Medications; Stress; Fatigue; Hormonal variations
28. What are the three major problems in behavioral observation studies?
Reactivity, Drift, and Expectancies
29. What is reactivity? How does performance change when people are not being observed or checked?
a. Observing the observers; it increases the accuracy and is done when people are coding actual behavior
30. What is drift? How does it relate to contrast effect? How can drift be addressed?
People tend to drift away from the protocol.
Contrast effect is the tendency to rate the same behavior differently when observations are repeated in the same context. It is one of the most common behaviors of drift. It can be addressed through repeated checks
32. How well do people do in detecting deception/lies?
Not very well
33. What is the Halo Effect? Be able to think of examples where the Halo Effect applies.
Attribute positive attributes independently of the observed behavior. Linked with general standoutishness-or attractiveness
34. What does a good interviewer know how to do?
Keep interview answers flowing
35. How are interviews similar to tests?
Both used to gather information about the participant
36. Define interpersonal influence and interpersonal attraction. How are they interrelated?
Interpersonal Influence: Degree to which one person can influence another
Interpersonal attraction: Degree to which people share a feeling of understanding, mutual respect, similarity, etc.
37. Define and note what types of statements should be avoided to elicit as much information as possible?
Judgmental statements:
Evaluative Statements: GOOD, BAD, excellent, terrible
Probing Statements: Like using Why?
Hostile Responses: directs anger toward the interviewee
False reassurance: reassuring subject when they shouldn’t be reassured.
38. What is the main goal in interviewing?
Gather information
39. Define transitional phrase. If it fails, what responses should be used to continue the theme? Define.
a. Once a response dies, the interviewer responds to keep the interaction flowing with minimum effort.
Transitional phrase (“Yes”, “I see”)
If it fails,
– Verbatim Playback – Repeats the exact words
– Paraphrasing and Restatement – Repeats the response using different words
– Summarizing – Pulls together the meaning of several responses
– Clarification Response – Clarifies the response
– Empathy & Understanding – Communicates interviewer knows how they feel
40. When should direct questions be used in an interview?
When you need to fill in important details; maybe at end of interview
41. What are the advantages and disadvantages of using structured clinical interviews?
a. Everyone gets same questions in same order
b. – Uses specified rules for probing so all interviewees are handled the same.
c. – They offer reliability, but sacrifice flexibility.
d. – Frequently used in research
Disadvantage:
Requires cooperation (particularly hard in some psychiatric and forensic settings).
– Relies exclusively on the respondent making the assumptions questionable.
42. What is the purpose of a mental status examination? What areas are typically covered?
It is given at the hospital and areas are psychosis, brain damage, neurologic, and mental status
43. Define general standoutishness. How does appearance play a role?
Attraction to subject you are interviewing
Take one characteristic and take it out on all their characteristics.
44. How much higher is interview reliability for structured interviews?
Double
45. What is a major criticism of structured interviews?
It sacrifices flexibility
46. What is social facilitation? Be prepared for examples.
We act like the people around us
47. What is the largest source of error in interviews?
JUDGING
48. What were the three independent research traditions identified by Taylor to study human intelligence?
Psychometrics, Information Processing, and Cognitive
49. Through what 3 facilities did Binet believe intelligence expressed itself? What two major concepts guided him?
Judgment, Attention, and Reasoning; guided by two principles: General Mental Ability and Age Differentiation
50. Know age differentiation, mental age, general mental ability, and positive manifold.
Age differentiation: deferentiating older from younger children by older’s greater capacity
Mental age: equivalent age capabilities of a child regardless of his/her chronological age: obtained through age differentiation
General Mental Ability: Total product of the various separate and distinct elements of intelligence: sum of the parts
Positive manifold: good in one, good in others
51. Binet searched for tasks that could be completed by what percentage of children in a particular age group?
66.67% to 75% (2/3 to ¾)
52. What concept did Spearman introduce? What does this concept mean?
The “g”: general mental ability, which he got from positive manifold.
53. What statistical method did Spearman develop to support his notion of g?
factor analysis
54. According to the gf-gc theory, what are the two basic types of intelligence? How do they differ?
Fluid and crystallized
55. Know and be able to calculate IQ using mental age and chronological age
IQ is a unit of measure for expressing the results of intelligence tests and it is a ratio score. It is found by dividing mental age/ chronological age.
56. What is a deviation IQ and how was it used in the Stanford-Binet scale?
Standard score with M=100; SD=15; used to compare age groups
Got rid of MA/CA
57. Define and differentiate basal and ceiling. Be able to give an example of each.
Basal: minimum number of correct responses obtained ex: go down an exam and say that person gets five correct in a row. Then they are at that level.
Ceiling: Number of wrong answers indicating items too difficult
Ex: go up an exam and see which items are too hard… keep on going until you get five incorrect.
58. What factors did Wechsler focus on that those before him had not?
Non-intellectual factors
59. What were some criticisms of the Binet scale by Wechsler?
No performance, it was designed for children, didn’t take into account non-intellectual factos, and was a one-score test.
60. What is the age range of the Wechsler scales?
Goes from 16 to age 90 and 11 months
61. Why is the inclusion of a point scale a significant improvement? What did a performance scale add?
Credits or points are assigned to each item

Individuals receive credit for each item passed
Allows analysis of individual in several areas
Allows you to sum together items together so you can score it in a similar ways

Performance scale added nonverbal measure of intelligence
Attempt to overcome cultural and language barriers and education.
62. Know which subtests are verbal and nonverbal. Know how they are divided into index scores
Four Index: Verbal Comprehension Index; Perceptual Organization Index; Working Memory Index; Processing Speed Index

Verbal subtests: Vocabulary, Similarities, Arithmetic, Digit Span, Information, Comprehension, Letter-number sequencing
Nonverbal subtests: Picture completion, Digit symbol-coding, Block Design, matrix Reasoning, Picture arrangement, symbol search, Object Assembly.

Verbal Comprehension Index: Has Vocabulary, Similarities and Information
Perceptual Organization Index: Picture Completion, Block Design, and Matrix Reasoning
Working Memory: Digit Span, Arithmetic, and Letter Number Sequencing
Processing Speed Index: Digit Symbol-Coding, and Symbol Search
63. Be able to know and differentiate the major functions measured by each subtest
Vocabulary: Vocabulary Level
Similarities: Abstract thinking
Arithmetic: Concentration
Digit Span: Immediate Memory, Anxiety
Information: Range of Knowledge
Comprehension: Judgment
Letter-number sequencing: Freedom from Distractibility
Picture Completion: Alertness to detail
Digit Symbol-Coding: Visual-Motor Functioning
Block Design: Nonverbal reasoning
Matrix Reasoning: Inductive reasoning
Picture Arrangement: Planning ability
Symbol Search: Information Processing Speed
Object Assembly: Analysis of Part-Whole relationships
64. Understand and be able to differentiate each WAIS subtest from the others
Vocab subtest: Ability to define words; most stable, and best single measure of intelligence
Similarities: Consists of paired items of increasing difficulty; most sensitive of verbal subtest to cerebral dysfunction
Arithmetic: Contains 15 relatively simple story problems; need to retain info in memory and manipulate it; concentration, motivation, and memory are the underlying main factors in performance
Digit Span: Requires subject to repeat digits forwards and backwards; measures short-term auditory memory; Non-intellectual factors easily influence this subtest: like anxiety, distraction, and inattention
Information: Items increase in difficulty; includes abilities to comprehend, follow directions and respond; VERY culturally specific
Comprehension: More a measure of social awareness; what should be done in this type of situation, provide a logical explanation of a rule or phenomenon; Define proverbs; emotional difficulties play a role
Letter-number sequencing: Asked to reorder lists of numbers and letters; related to working memory and attention
Picture Completion: Subject is shown a picture with something missing
Digit Symbol-Coding: Requires subjects to copy symbols and code numbers with the symbols in the box.
Block Design: Uses 9 colored blocks and a stimulus booklet and reproduce design; design increases in difficulty; Block design and Matrix reasoning are most sensitive in Performance IQ to cerebral dysfunctions.
Matrix Reasoning: Subject presented with a nonverbal figure and stimuli and must find the relationship between the two; good measure of abstract reasoning and information processing
Picture Arrangement: Requires subject to notice details and cause and effect relations. Have pictures and are asked to put the pictures into order.
Symbol Search: Ability to rapidly scan and process info; this shows the role of speed of information-processing in intelligence.
Object Assembly: Subject assembled to put together the puzzle quickly as possible; requires visual analysis and its coordination with simple assembly skill.
65. What are the mean, standard deviation, and range for scaled scores, standard scores, and index scores?
Scaled Scores: M=10, SD=3, Range=1-195
Standard Scores: M=100, SD=15, Range=45-155
Index Scores: M=100, SD=15, Range=50-150
How is IQ calculated
a. Verbal IQ (VIQ): The five raw verbal subtests scores are converted to age-corrected scaled scores which are summed. They are then compared with the standardization sample.
b. • Performance IQ (PIQ): The five raw performance subtests scores are converted to age-corrected scaled scores which are summed. They are then compared with the standardization sample.
c. • Full-Scale IQ (FSIQ): Follows the same principles of VIQ and PIQ. Obtained by summing the age-corrected scaled scores of the verbal and performance subtests and comparing the subject to the standardization sample.
67. Name the index scores and what the purpose of each index is.
a. Verbal Comprehension Index
i. Measures acquired knowledge and verbal reasoning; measure of crystallized intelligence
b. Perceptual Organization Index
i. Measure of Fluid Intelligence; Attention to detail and visual motor integration influence this
c. Processing Speed Index
i. How quickly your mind works
d. Working Memory Index
i. Information we actively hold in our minds, NOT stored information
68. What is pattern analysis? What are the concerns when using such a method?
When one looks for large differences in subtest scaled scores; Results of diseases using pattern analysis were inconclusive and contradictory; Doesn’t take into account individual variability well; Need to proceed WITH EXTREME CAUTION
69. What was one of the major changes of the WISC-IV from previous versions?
No PIQ and VIQ. Just 4 index scores
70. What is a hold subtest?
Intelligence in that test is retained after brain damage
71. Which verbal and performance subtests are most sensitive to cerebral dysfunction? Which are considered “hold” subtests?
Most sensitive are the Similarities, Block-design, and Matrix reasoning; Hold subtests: Vocab
72. Where do traditional intelligence tests fail in the study “normal” abilities?
In extreme populations; those with physical, sensory, mental, or social disabilities
73. What are the disadvantages of alternative intelligence tests when compared to Binet and Wechsler?
Weaker psychometrics; less validity and documentation; weaker standardized sample; Test manual instructions; less stable

Scores are NOT interchangeable with Binet and Wechsler and specificity limits what abilities the test measures.
74. What are the advantages of alternative intelligence tests when compared to Binet and Wechsler?
It hits extremes, a specialized test, not as culturally dependent, and not quite crystallized
75. What theme in relation to future intelligence do you notice about infant development tests?
They do NOT predict future intelligence; only can tell with developmental features
76. Compare and contrast surveillance and screening. Be able to identify either.
Surveillance: Survey wide screen; intervene ONLY when you see those who aren’t developing.
Screening: You look for a specific area of concern… You NEED additional evaluation more monitoring
77. Which two infant development tests were discussed in class? What are some disadvantages of each of these tests?
Bayley: Disadvantages are there are that the psychometrics breakdown at lower age groups. IT does NOT predict later IQ, BUT extreme low scores predict MR later in life.
Brazelton Neonatal Assessment Scale: Disadvantages are that there aren’t any norms and has poor test reliability. Also, does NOT predict future IQ.
78. Know sensitivity, specificity, false positives, true negatives, etc.
Sensitivity: Accuracy of test finding developmental delays (Find True positives)
Specificity: Accuracy of test finding those who are NOT delayed (Find False Negatives)
False Positives: Is not sick but is diagnosed
True Negatives: Is not sick, is not diagnosed
False Negatives: Is sick, NOT diagnosed
True Positives: Is sick, is diagnosed
IF you are testing sensitivity, you will get many false positives: goal to get all diseased ones
If you are testing for specificity, you will only get true positives, and you hence have false negatives.
79. What are acceptable sensitivity and specificity levels for developmental screening tests?
70-80%
80. How is a learning disability currently defined in the school systems? Is this a good method? Why or why not?
Learning disability is 1.5 standard deviations of IQ and achievement …and NO not good method…
81. For what was the Woodcock Johnson-III designed? (i.e., what does it assess?)
Achievement (academic performance):
Woodcock-Johnson (WJ-III)
• Cognitive (IQ) and achievement batteries
– Most testers use only achievement
• Great psychometrics and validity research
• Used to define Learning Disabilities, based on
discrepancy from IQ