• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/75

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

75 Cards in this Set

  • Front
  • Back
reliability
the degree of dependability, consistency, or stability of scores on a measure used in selection research
Predictions concerning selection must be based on the results of “true measures” of differences that have:
Consistency in results
Dependability of measurement
Stability over time
Errors of Measurement
Factors that affect obtained scores but are not related to the characteristic, trait, or attribute being measured
Obtained Score Components:
Xobtained = Xtrue + Xerror
Xobtained =
obtained score for a person on a measure
Xtrue =
true score on the measure, that is, actual amount of the attribute measured that a person really possesses
Xerror =
error score on the measure that is assumed to represent random fluctuations or chance factors.
True Score
The mean or average score made by an individual on many different administrations of a measure if external and inter conditions were perfect
Error Score
The score made by individuals that is influenced by factors present at the time of measurement that distort individuals’ scores either over or under what they would have been on another measurement occasion
Reliability Coefficient
An index that summarizes the estimated relationship between two sets of measures
The higher the coefficient, the higher the reliability estimate and the lower the error of measureme
Principle Methods of Estimating Reliability
Test-retest or stability
Parallel or equivalent forms
Internal consistency
Interrater reliability estimates
Test-Retest Reliability
The same measure is used to collect data from the same respondents at two different points in time
Factors that may affect test-retest reliability
Memory (overestimate)
Learning (underestimate)
Use test-retest reliability when
-the length of time between the two administrations is long enough to offset the effects of memory or practice

-there is little reason to believe that memory will affect responses to a measure

-it can be determined that nothing has occurred between the two testings that will affect responses

-information is available on only a single item measure
Internal Consistency Reliability Estimate
An index of a measure that shows the extent to which all parts of a measure are similar in what they measure
Internal Consistency Procedures
Split-half reliability
Kuder-Richardson reliability
Cronbach’s coefficient alpha (α) reliability
Sources of Measurement Error
What is being rated (e.g., employee behavior)
Who is doing the rating (rater characteristics)
Interrater Reliability Estimates
Sources of Measurement Error

Consistency of the implementation of a rating system

Procedures for Calculating Interrater Reliability
Interrater Agreement
Percentage of rater agreement
Kendall’s coefficient of concordance (W)
Cohen’s kappa (K)
Reliability Analyses
help determine the dependability of data we will use in selection decision making

help in estimating the amount of error included in scores on any measure we choose to study
What Does a Reliability Coefficient Mean?
The extent (in percentage terms) to which individual differences in scores on a measure are due to “true” differences in the attribute measured and the extent to which they are due to chance errors
Characteristics of Reliability Coefficients
They are specific to the reliability estimation method and group on which they are calculated

They are a necessary but not a sufficient condition for validity

They are based on responses from a group of individuals

They are expressed by degree, and are determined ultimately by judgment
Factors Affecting Estimated Reliability
Method of estimating reliability
Stability
Sample
Individual differences among respondents
Length of a measure
Test question difficulty
Homogeneity of content response format
Administration and scoring
Validity: A Definition
The degree to which available evidence supports inferences made from scores on selection measures.
Importance of Validity in Selection
Shows how well a predictor (such as a test) is related to important job performance criteria

Can indicate what types of inferences may be made from scores obtained on the measurement device
Validation Study
Provides evidence for determining the accuracy of judgments made from scores on a predictor about important job behaviors as represented by a criterion.
Types of Validation Strategies
Content validation
Criterion-related validation
Concurrent and predictive strategies
Construct validation
Content Validity
Is shown when the content (items, questions, etc.) of a selection measure representatively samples the content of the job for which the measure will be used.
Why Content Validation?
Is applicable to hiring situations involving a small number of applicants for a position

Focuses on job content, not job success criteria

Increases applicant perception of the fairness of selection procedures
“Content of the job”
Job behaviors, the associated knowledge, skills, abilities, and other personal characteristics (KSAs) that are necessary for effective job performance
Job Content Domain
The behaviors and KSAs that compose the content of the job to be assessed

Content of the measure must be representative of the job content domain for the measure to possess content validity
Content Validation Strategy
Involves construction of a new measure rather than the validation of an existing one.

Emphasizes the role of expert judgment in determining the validity of a measure rather than relying on statistical methods
Face Validity
Concerns the appearance to job applicants of whether a measure is measuring what is intended—the appearance to applicants taking the test that the test is related to the job.

Increases acceptability of a measure
Major Aspects of Content Validation
Conduct a Comprehensive Job Analysis

Selection of Experts Participating in Study

Specify Selection Measure Content

Assessment of Selection Measure and Job Content Relevance
Content validation is not appropriate when:
Mental processes, psychological constructs, or personality traits are not directly observable but inferred from the selection device.

The selection procedure involves KSAs an employee is expected to learn on the job.

The content of the selection device does not resemble a work behavior; when the setting and administration of the selection procedure does not resemble the work setting
Content validity
Focuses on the selection measure itself

Is based on a broader base of data and inference

Is generally characterized as using broader, more judgmental descriptors (description)
Criterion-related validity
Focuses on an external variable
Is narrowly based on a specific set of data
Is couched in terms of precise quantitative indices (prediction)
Concurrent Validation Strategy
Both predictor and criterion data is obtained on a current group of employees, and statistical procedures are used to test for a statistically significant relationship (correlation coefficient) between these two sources of data

Sometimes referred to as the “present employee method” because data is collected for a current group of employees
Concurrent Validation
Conduct analyses of the job.

Determine relevant KSAs and other characteristics required to perform the job successfully.

Choose or develop the experimental predictors of these KSAs.

Select criteria of job success.

Administer predictors to current employees and collect criterion data.

Analyze predictor and criterion data relationships
Predictive Validation
Conduct analyses of the job.

Determine relevant KSAs and other characteristics required to perform the job successfully.

Choose or develop the experimental predictors of these KSAs.

Select criteria of job success.

Administer predictors to job applicants and file results.

After passage of a suitable period of time, collect criterion data.

Analyze predictor and criterion data relationships.
Construct
A postulated concept, attribute, characteristic, or quality thought to be assessed by a measure.
Construct Validation
A research process involving the collection of evidence used to test hypotheses about relationships between measures and their constructs.
What precisely does a statistically significant coefficient mean
Coefficient of determination

Expectancy tables and charts

Utility analysis
Coefficient of determination
The percentage of variance in the criterion that can be explained by variance associated with the predictor.
Utility analysis
An economic interpretation to the meaning of a validity coefficient
Linear Regression
The determination of how changes in criterion scores are functionally related to changes in predictor scores
Regression Equation
Mathematically describes the functional relationship between the predictor and criterion
-Simple regression assumes only one predictor
-Multiple regression assumes two or more predictors
Utility Analysis: A Definition
Shows the degree to which use of a selection measure improves the quality of individuals selected versus what would have happened if the measure had not been used.
Validity Generalization
Uses evidence accumulated from multiple validation studies to show the extent to which a predictor that is valid in one setting is valid in other, similar settings
Situational Specificity Hypothesis (Ghiselli)
The validity of a test is specific to the job or situation where the validation study was completed.
Deficiencies Affecting Validity
The use of small sample sizes (sampling error)
Differences in test or predictor reliability
Differences in criterion reliability
Differences in the restriction of range of scores
Criterion contamination and deficiency
Computational and typographical errors
Slight differences among tests thought to be measuring the same attributes or constructs
Conclusions
It is not necessary to conduct validity studies within each organization for every job

Mental ability tests can be expected to predict job performance in most, if not all, employment situations
job performance measures
to serve as criteria measures in validation studies
production data
hr personnel data
training proficiency
judgmental data
types of job performance measures
Types of Production Data
Quantity—usually expressed as the number of units produced within a specified time period.
Quality—the goodness of the product indirectly measured by defects, errors, or mistakes identified either per number of units or per unit of time.
The assumption that promotions are the result of high performance rather than other factors such as job tenure and organization change/growth.
Limitations imposed by short career ladders.
Distinguishing lateral from vertical promotions.
Accounting for promotions voluntarily declined
Issues in Measuring Promotions
Training Proficiency
A measure of employees’ performance immediately after completing a training program.
Promotion
A measure of career progress, based on the number of vertical job changes for an individual that result in increased responsibility and authority
Measuring Proficiency
Standardization of training programs—instructional materials, techniques, training mastery measures
Linking of training effects to increases in worker KSAs
Measures of Training Proficiency
Training instructor judgments about trainees
Scores on paper-and-pencil tests
Scores on work sample tests
problem of bias
avoiding supervisor biases in making judgment
training supervisors to avoid common errors
systematically collecting data to ensure complete data
Halo—
rating the subordinate equally on different performance scales because of a rater’s general impression of the worker.
Leniency/Severity—
rating a disproportionate number of workers either high or low respectively.
Central tendency—
rating a large number of subordinates in the middle of the scale.
Signs
Selection tests that are indirect indicators of an individual’s predisposition to behave in certain ways.
Samples
Selection tests that gather information about behaviors that are consistent with the job behaviors being predicted.
Motor Tests
Require the physical manipulation of things—operating a machine, installing a piece of equipment, or making a product.
Verbal Tests
Require the use of spoken or written messages or interpersonal interaction if the problem situation is primarily language or people oriented.
Trainability Tests
Are most often used for jobs that do not presently exist and specialized jobs for which extensive training is necessary.
Establish Independent Test Sections
Develop the test such that an applicant’s performance on one part of the test is not closely tied to another part of the test.
Assessment Center (AC)
A procedure for measuring KSAs in groups of individuals (usually 12 to 24) that uses a series of devices (exercises), many of which are verbal performance tests.
Uses of ACs
Selection—identifying participants who demonstrate behaviors necessary for the position considered.
Career development—determining those behaviors each participant does well and those in which each is deficient.
In-Basket
A paper-and-pencil test designed to replicate administrative tasks of the job under consideration.
Leaderless Group Discussion (LGD)
A small group of applicants are each assigned roles and charged with resolving either a competitive or a cooperative scenario.
Situation Judgment Tests (SJTs)
“Low-fidelity” simulations that present a series of written descriptions of work situations and multiple responses to each situation.