Professional Documents
Culture Documents
1990,43
DIANA L. DEADRICK
Owen Graduate School of Management
Vanderbilt University
ROBERT M. MADIGAN
Management Department
Virginia Polytechnic Institute and State University
The concept of dynamic criteria has been the subject of a recent de-
bate regarding both the definition and prevalence of the phenomenon
(Austin, Humphreys, & H u h , 1989; Barrett & Alexander, 1989; Bar-
rett, Caldwell, & Alexander, 1985). The present paper questions the
adequacy of the conceptual framework underlying the debate and pro-
vides data supporting a refined concept of dynamic criteria. The in-
cidence and possible causes of change in relative performance were
investigated using weekly performance data from 509 sewing machine
operators. Analyseswere conducted to determine the degree of perfor-
mance consistency, potential moderators of consistency, and the sta-
bility of predictor-criteria relationships using multiple predictors and
criteria. Results revealed a steady decline in performance stability co-
efficients as the interval between measures increased. This decay was
evident regardless of employees’ prior job experience, cognitive abil-
ity, or psychomotor ability. Analyses of predictive validity coefficients
revealed temporal changes in validity for both objective and subjective
criteria,but not in the expected direction. The validity of cognitive abil-
ity increased, the validity of psychomotor ability was stable, and that of
prior job experience decreased over time. Implications for theory and
research are discussed.
Support for this paper was provided to the first author by the Dean’s Research Fund,
Owen Graduate School of Management. The authors contributed equally to this article.
We wish to thank Tom Mahoney, Rich Oliver, and Cliff Ball of the Owen Graduate
School of Management, Vanderbilt University, and three anonymous reviewers for their
helpful comments.
Correspondence and requests for reprints should be addressed to Diana L. Deadrick,
Owen Graduate School of Management, Vanderbilt University, 401 21st Avenue South,
Nashville, TN 37203.
COPYRIGHT@ 1990 PERSONNEL PSYCHOLOGY, INC
717
718 PERSONNEL PSYCHOLOGY
Concept Definition
and operationally weak. Their debate centers on the intrinsic and sys-
tematic nature of change in predictor-criterion correlations over time
(definition #2) and on the prevalence and systematicnature of change in
relative performance over time (definition #3). Barrett et al. (1985) ini-
tially argued that although changes in criterion rank-ordering might oc-
cur, this definition of dynamic criteria “. ..assumes practical importance
only in relation to potentially consequent changes in validity” (pp. 52-
53). Based on their reviews and reanalyses of prior validation studies
incorporating repeated criterion measures, Barrett et al. (1985, 1989)
found little support for this concept of dynamic criteria and, therefore,
implicitly advocate changes in validity coefficients (definition #2) as the
most meaningful evidence of dynamic criteria. More recently, Barrett
and Alexander (1989) criticized the assertion that the simplex pattern
is a ubiquitous phenomenon with regard to either definition of dynamic
criteria, arguing that the simplex approach relies on a methodological
versus a theoretical basis for predicting changes.
In contrast to Barrett et al. (1985, 1989), Austin et al. (1989) ap-
pear to be concerned primarily with developing a better understanding
of criteria, per se, and focus on fluctuations in the rank-ordering of cri-
terion scores over time (definition #3) as the “core of the dynamic cri-
terion concept” (p. 589). Although they cite supporting evidence in the
form of both validity and performance stability coefficients, they empha-
size that the essence of dynamic criteria is represented by the decreas-
ing stability in performance intercorrelations over distant time periods
(p. 592). We take the position here that neither of these operational
definitions provides a sufficient basis upon which to advance our under-
standing of dynamic criteria because they do not indicate whether the
temporal changes reflect actual changes in job performance or changes
in the performance evaluation context. This confounding of different
types (sources) of change is reflected in the existing literature and cur-
rent debate. In this study we differentiate between criterion changes
attributable to individual differences (perfiormanceconsistency), changes
attributable to the organizational context (evaluation consistency), and
changes attributable to the measurement procedure (measurement re-
&ability). It is our contention that this distinction is necessary for any
meaningful discussion of the concept of dynamic criteria.
Performance consistency, as defined here, refers to the systematic
changes in critical job behaviors or outcomes over time that are at-
tributable to individual differences (Kane, 1982; Wernimont and Camp-
bell, 1968); therefore, the ultimate sources of criterion variability are
within the person. This concept is implied in Ghiselli’s (1956) discussion
of dynamic criteria, in which he suggests that there may be a uniform
pattern of change in performance over an extended period of time as
720 PERSONNEL PSYCHOLOGY
workers learn and develop on the job. Performance consistency has been
investigated by Rothe (1978) and by Rambo, Chomiak, and Price (1983)
who analyzed repeated measures of output obtained under conditions
of context stability (motivation) and/or task stability. Performance con-
sistency could also be ascertained via validity coefficients in situations
where the ability requirements of the job/task remain stable while indi-
viduals’ abilities change over time. This possibility has been referred to
as the changing-subject model (Alvares & H u h , 1972; Henry & H u h ,
1987).
Evaluation consistency, as defined here, refers to the temporal con-
sistency of the performance evaluation system. Criterion stability can
be affected by changes in the organization’s objectives, performance re-
quirements, job design, and other factors that might cause a change in
the relative importance of performance dimensions, thus altering the
composition of global (summary) criteria. In essence, performance is
implicitly redefined, conceptually and/or operationally; hence, the im-
petus for change is in the work environment and/or evaluation system.
The employee’s actual performance behavior or outcomes could be con-
sistent over time, but the overall performance evaluation would suggest
otherwise. This notion of dynamic criteria is evident in Prien’s descrip-
tion of dynamic criteria and organization change (Prien, 1966): The tran-
sitional nature of organizational needdobjectives results in shifting per-
formance expectations and requirements, although job duties might re-
main static. Evaluation consistency is also implied in Alvares and Hulin’s
(1972) changing-task model, in which the task structure and ability re-
quirements change over time while individuals’abilities remain constant,
and in Murphy’s (1989) dynamic model of job performance, in which
structural changes occur in the job and/or work environment. Although
evaluation consistency has not been directly studied in the context of dy-
namic criteria, Barrett and Alexander (1989) refer to this phenomenon
as the vnderlying cause of fluctuating predictor-criterion validities in
their rdview of dynamic criteria from educational psychology. Citing
opportunity bias, differential performance standards, and information-
processing differences in tasks, Barrett and Alexander (1989) concluded
that predictor-criteria relationships may decline, remain the same, or in-
crease over time, depending on the situation.
The concept of evaluation consistency should be distinguished from
measurement reliability. Even if the performance evaluation system and
work context remain stable, the temporal stability of observed criterion
scores and/or validity coefficients could still be affected by measurement
error. However, while the distinction between evaluation consistency
and measurement reliability is conceptually straightforward, it is often
difficult or impossible to operationally distinguish between them. For
DEADRICK AND MADIGAN 721
Rambo et al., 1983; Wernimont & Campbell, 1968). The focus on re-
curring critical tasks provides a control for work factors that might affect
stability but are characteristics of the situation rather than the individ-
ual. Third, in order to control other personal and contextual factors, the
sample should be an intact group of employees working on the same job
and under common performance standards and operating procedures.
Although temporal stability of predictor-criterion relationships is not
a required condition for evidence of performance consistency, such anal-
yses can and should be employed to examine possible causes of stabil-
itylchange (Ackerman, 1989; Murphy, 1989). Motivational andlor per-
sonal (ability) factors that might explain temporal changes in perfor-
mance should therefore be incorporated in the study design (Ackerman,
1989; Austin et al., 1989; Barrett & Alexander, 1989; Henry & Hulin,
1987; Rambo et al., 1983; Wernimont & Campbell, 1968).
Prior Research
A third limitation of the Rothe (1978) and Rambo et al. (1983) stud-
ies is that they considered only an objective criterion; the more com-
monly used subjective performance measures were not included in their
analyses of consistency. Rothe noted that subjective data might have
contributed additional information, but that it “would also have con-
tributed controversy because of the lack of objectivity” (p. 46). How-
ever, an assessment of performance consistency need not preclude sub-
jective performance measures. The only stipulation is that the perfor-
mance ratings pertain to specific dimensions of performance behaviors
or outcomes (Wernimont & Campbell, 1968). Moreover, the use of both
objective and subjective (repeated) measures of specificperformance di-
mensions would provide some insight into the related issue of criterion
(rating) validity. Consider a case where a test-retest correlation between
dimensional ratings is high. Such rating consistency does not necessar-
ily indicate performance consistency. Rater error (i.e., first-impression,
halo) can produce rating consistency that is independent of actual (ob-
jective) performance consistency. In this case, the high reliability esti-
mate would actually reflect systematic error. A more informative ap-
proach is to examine the degree to which ratings of a specific perfor-
mance dimension, taken at different points in time, correlate with ob-
jective measures of the same performance dimension covering the same
period of time. Given high convergence among these different methods
of measurement, one would expect any changes in validity over time to
be similar for these matched criteria (see Smith, 1976).
Method
Participants
Job performance data and ability measures were collected for sewing
machine operators employed at five nonunionized garment manufac-
turing plants in the Southeast. All five sewing plants were owned by
the same company, produced the same kind of garments using similar
equipment and operating procedures, and operated under a uniform set
of management policies, procedures, and record-keeping. The sample
sizes varied across plants (80-338); however, the demographic and abil-
ity characteristicswere quite similar. Because of the high degree of stan-
dardization of operations and operator similarity across plants, analyses
reported here were conducted on a combined sample.
Sewing machine operators performed a single operation during a
shift and were paid on a piece-rate basis. Although jobs were laid out
in a “production line,” each operator independently stitched bundles of
garment pieces from a large in-process inventory. As a result, the pro-
duction performance of any one operator was not dependent on the per-
formance of other operators. The piece-rate standards were determined
by industrial engineering studies typical of the industry. Thus, within the
limits of error of those studies, jobs were equated along a common pay
scale and differences in piece-rate (production) earnings reflect differ-
ences in production performance.
Operators’ work assignments (specificsewing operations) were made
by supervisors based on the plant production schedule as opposed to
either seniority or ability qualifications. As in all such garment plants,
there undoubtedly were some sewing operations that offered better earn-
ing opportunities, but there was no formal system to accommodate op-
erator preferences. Earnings temporarily decreased in some instances
when new models were introduced. However, such changes typically af-
fected all operators, and earnings levels quickly moved back to previous
levels.
The initial sample consisted of 932 operators hired during a 10-
month period. However, due to turnover and/or missing data, three sub-
samples were used for the analyses. The first subsample (N = 509) was
composed of employees for whom we had objective performance data
for at least the first 6 months on the job. This sample was used to ana-
lyze the first research proposition regarding the temporal decay in per-
formance consistency. Participants were all female, of which 67% were
white and 33% were black. The average age was 26.4 years old (range =
17-59) and the average amount of previous sewing experience was 12.9
months. Sixty-five percent of the participants had no previous sewing
728 PERSONNEL PSYCHOLOGY
Procedure
Results
Pe$omzance Characteristics
TABLE 1
Means and Standard Deviations of Pei$ormance and Aptitude hriables
for the Three Samples
Sample 1 Sample 2 Sample 3
output M SD M SD M SD
Variablesa
Week 1 $ 2.07 1.17 $1.98 1.02 $2.06 1.08
Week 2 2.42 1.23 2.35 1.13 2.44 1.15
Week 3 2.67 1.27 2.62 1.20 2.70 1.23
Week 4 2.89 1.22 2.84 1.14 2.90 1.09
Week 8 3.49 1.24 3.43 1.16 3.47 1.18
Week 12 3.88 1.25 3.84 1.16 3.87 1.15
Week 16 4.12 1.21 4.08 1.10 4.11 1.07
Week 20 4.32 1.21 4.28 1.11 4.34 1.06
Week 24 4.46 1.19 4.47 1.18 4.51 1.12
Month 1 $ 2.51 1.17 $2.45 1.07 $2.53 1.08
Month 2 3.29 1.21 3.25 1.12 3.31 1.11
Month 3 3.74 1.20 3.70 1.12 3.72 1.11
Month 4 4.01 1.15 3.98 1.06 4.01 1.04
Month 5 4.24 1.16 4.21 1.08 4.27 1.03
Month 6 4.40 1.15 4.41 1.12 4.45 1.06
Months 1-3 $3.18 1.15 $3.13 1.06 $3.18 1.06
Months 4-6 4.22 1.11 4.19 1.04 4.24 1.00
Months 1-6 3.70 1.09 3.66 1.01 3.71 0.98
Rating variablesb
Quantity rating #1 2.5 1.18 2.5 1.15 2.4 1.16
Quantity rating #2 2.6 1.08 2.6 1.05 2.7 1.05
Global rating #1 3.2 1.04 3.2 1.00 3.2 1.01
Global rating #2 3.2 0.86 3.3 0.86 3.2 0.88
Aptitude variables'
Cognitive ability 272.1 34.41 272.5 34.05 276.6 33.20
Psychomotor ability 324.4 45.44 324.5 45.15 330.3 38.79
Job Family 5 composited 53.2 % 25.73 53.3 % 25.51 55.1 % 22.83
a N = 509 (Sample l), 413 (Sample 2), and 224 (Sample 3).
N = 294 (Sample l), 239 (Sample 2), and 224 (Sample 3) for the initial ratings. N = 487
(Sample l), 395 (Sample 2), and 224 (Sample 3) for the follow-up ratings. The Quantity
ratings were based on 5-point rating scales; the Global ratings were based on 6-point rating
scales.
'N = 419 (Sample l), 413 (Sample 2), and 224 (Sample 3).
The Job Family 5 composite score was obtained from the GATB and is a norm-
referenced score composed of psychomotor ability (56%) and cognitive ability (44%).
Consistency of Peflomance
TABLE 2
Median CorrelationsBetween Average Weekly Output at Different
Time-LaggedIntervals (K) for the Total Sample and for the Ability Subgroups
Total
Median r .92 .87 .81 .75 .71 .66 .55
Range .04 .04 .08 .07 .08 .08 -
Prior experience
Median T 91 .86 .80 .74 .69 .62 .55
Range .07 .09 .10 .10 .09 .09 -
No experience
Median r .90 .85 .76 .68 .65 .6 1 SO
Range .06 .16 .15 .17 .12 .13 -
High cognitive
Median r .89 33 .76 .67 .63 .57 .47
Range .09 .12 .10 .07 .06 .12 -
Low cognitive
Median T .91 .85 .74 .69 .66 .61 .55
Range .12 .10 .14 .22 .22 .16 -
High psychomotor
Median r .91 37 .80 .71 .67 .66 .59
Range .09 .10 .14 .17 .08 .12 -
Low psychomotor
Median T .92 .88 .80 .75 .72 .67 .52
Range .05 .09 .ll .21 .17 .19 -
Note: The values of “K” refer to the number of intervening weeks. The number of
correlation coefficients at each value of K is: 23 at K = 1week; 21 at K = 3 weeks, 17 at
K = 7 weeks; 13 at K = 11weeks; 9 at K = 15 weeks, 5 at K = 19 weeks; and 1 at K =
23 weeks. The range reported here refers to the difference between the highest and lowest
observed correlations for each time interval and sample.
a N = 509. The sample sizes for the Experienced and Inexperienced subgroups are 138
and 275, respectively;for High and Low cognitive ability are 101and 102, respectively; and
for High and L o w psychomotor ability are both 102.
(experience and aptitude). However, these data do not describe the na-
ture of the instability, nor do they indicate whether there are significant
differences in consistency across ability subgroups. Regression analyses
were employed as a means of examining the rate of change in stability
as a function of time separations. For this analysis, average correlations
were computed for each time interval K (via Fisher’s r-to-2 transforma-
tion and retransformation), and these averages were then regressed on
K. Estimates of Root Mean Square Error (RMSE) and R2were com-
puted in order to determine model fit. The regression analyses were con-
ducted on the total sample and the experience and aptitude subgroups.
The plot of average correlations on K revealed that a linear trend
(decline) fit the data well, a finding that sharply contrasts with the Rambo
et al. (1983) study in which a hyperbolic function best described their
data. A possible explanation for this difference is that the 6-month pe-
riod of the current study was too brief to identify the long term trend
revealed in the Rambo et al. study which covered 42 months. However
the data in Table 3 clearly suggest that the linear model is the best fit,
whereas Rambo et al. found that half of the total decrease in the mean
correlation occurred by week 18 and that the curvilinear trend was clearly
evident prior to 6 months. The ordinary least squares estimates for the
total sample presented in Table 3 revealed a significant decline in sta-
bility as the time interval increased (slope = -.01, p < .0001) and a
good model fit (RMSE = .01 and R2 = .99). However, a plot of resid-
uals on K revealed cyclical trends in the error terms, which is indicative
of first-order autocorrelation (i.e., the adjacent residuals were not inde-
pendent). The test for autocorrelation was significant (Durbin-Watson’s
d = .60, p < .Ol), thus necessitating a time-series model that could ac-
count for the autocorrelation and therefore improve the fit of the model
and the reliability of the model estimates. These results, shown in the
lower half of Table 3, were quite simikr to the ordinary least squares
estimates for the rate of change in stability (slope = -.Ol), the model
error (JXMSE = .Ol), and the proportion of variance in mean stability
that was accounted for by time separations (R2= .99). On the basis of
these results, our proposition 1 was supported: There was a significant
decline in performance consistencyas the interval between performance
measurement occasions increased.
Table 3 also presents the regression results for the experience and
aptitude subgroups, thus providing evidence pertaining to propositions
2 and 3. With regard to the impact of experience on performance con-
sistency, the plot of average correlations on K revealed a linear trend
for both groups, and the initial regression analysis indicated a slightly
stronger decline in stability for the inexperienced subgroup. However,
due to significant autocorrelation in both subgroups, autoregressive time
DEADRICK AND MADIGAN 735
TABLE 3
Regression Model Estimates of Average Performance Consistency
Over Time Intervals (K)for the Total Sample and for the Ability Subgroups
No ex- Priorex- Low High Lowpsy- High psy-
Sample Total perience perience cognitive cognitive chomotor chomotor
OLS Estimatesa
Intercept .91 .88 .90 .87 .89 .92 .89
Slope -.01 -.02 -.01 -.01 -.02 -.01 -.01
RMSE .01 .02 .01 .02 .01 .02 .01
R2 .99 .98 .99 .95 .99 .96 .98
ARl Estimatesb
Intercept .92 .89 .91 .89 .89 .93 .90
Slope -.01 -.02 -.02 -.02 -.02 -.02 -.01
RMSE .01 .01 .01 .01 .01 .01 .01
R2 .99 .99 .99 .99 .99 .98 .99
Note: K = 23 time intervals
a OLS refers to OrdinaryLeast Squares estimates. The model estimates were significant
at p < .OOOl.
ARl refers to first-order Autoregressive estimates. The model estimates were signifi-
cant at p < .0001.
series models were computed (lower half of Table 3), which revealed
equivalent estimates of the decline in stability over K and equivalent
estimates of model fit. In addition, although the mean level of per-
formance consistency across K was slightly higher for the experienced
subgroup, this difference was not statistically significant (t = 1.289, ns).
Thus, these results did not support our research proposition 2: Perfor-
mance consistency was not significantly higher for the experienced em-
ployees, and there was a significant decline in stability for experienced
as well as inexperienced employees.
The regression analyses of the cognitive and psychomotor ability sub-
groups produced similar results. Performance stability declined signifi-
cantly across K for the high and low subgroups, and there was no sig-
nificant difference between high and low groups in the mean level of
stability across K for either cognitive or psychomotor ability (t = -.83
and -.44, respectively). Hence, our research proposition 3 was not sup-
ported: Both high and low aptitude employees were characterized by
similar and significant declines in stability over time.
Consistency of Validity
TABLE 4
CorrelationsAmong Output Criteria and Predictors Over Time
1 2 3 4 5 6 7 8
-
Criterion variablesa
1. Month 1 average .91 34 .75 .69 .66 .95 .73
2. Month 2 average - .92 32 .76 .72 .98 .80
3. Month 3 average - .91 .83 .78 .96 .87
4. Month 4 average - .92 .83 .86 .95
5. Month 5 average - .91 .79 .98
6. Month 6 average - .75 .95
7. Months 1-3 average .83
8. Months 4-6 average
Predictive validityb
Prior experience .27 .23 .21 .13 .15 .16 .25 .15
Psychomotor ability .16 .17 .18 .16 .18 .20 .17 .19
Cognitive ability .09 .09 .10 .12 .17 .16 .09 .16
Note: N = 413.
a All criteria intercorrelations were significant at p < .OOO1
Validity coefficients were significant as follows: T = .10 to .11, p < .05; T = .12 to .18,
p < .01; r > .18, p < .OOOl.
Rating Validity
Discussion
TABLE 5
Correlations Between Objective-Subjective Criteria and
Predictors-CriteriaOver Time
1 2 3 4 5 6
Variablesa
1. Global rating #1 .53 .59 NA .42 NA
2. Global rating #2 .44 .62 .31 .41
3. Quantity rating #1 .50 .61 NA
4. Quantity rating #2 .49 .65
5. Months 1-3 average .80
6. Months 4-6 average
Predictive validityb
Prior experience .19 .18 .21 .I7 .37 .28
Cognitive ability .28 .27 .15 .17 .07 .17
Psychomotor ability .18 .15 .13 .13 .ll .13
Note: N = 224. The cells denoted “NA”reflect illogical relationshipsdue to the temporal
nature of the ratings.
a All criteria intercorrelations were significant at p < .WI.
Predictive validity coefficients were significant as follows: T = .13 - .16, p < .05; T =
.17 - .24, p< .01; T > .24, p < .W1.
(model) changes did occur periodically in each of the plants. To the de-
gree the industrial engineering system did not properly compensate for
these changes, the relative earnings could have been affected. However,
such changes typically affected the entire production line, resulting in
a brief general decrease in production as the workers adjusted to new
sea-ingoperations. Moreover, these changes occurred at irregular inter-
vals. Hence it is unlikely that they were a primary determinant of the
steady decline in stability coefficients. Other, more subtle, situational
factors might have differentially affected the operators, but in our ex-
tensive exposure to the plants and workers we did not identify them. In
short, we believe both the context and the findings here point to individ-
ual level factors as the primary determinants of performance inconsis-
tency.
The research reported here, together with a growing body of evi-
dence for instability in performance and validity (see Murphy, 1989),
suggests several research avenues that need to be explored further. First,
more longitudinal research needs to be conducted that examines re-
peated measures of both performance and ability. Repeated measures of
global and specific performance criteria will provide evidence for crite-
rion validity and equivalence (Binning & Barrett, 1989; Ironson, Smith,
Brannick, Gibson & Paul, 1989;James, 1973; Smith, 1976). Ideally, mul-
tiple methods of performance measurement would be utilized to better
our understanding of the “conceptual criterion.” Repeated measures of
ability will enable us to make more informed assumptions about the rel-
ative stability of ability factors as well as the stability of predictive validity
( H u h et al., 1990; Murphy, 1989).
Second, the research on performance consistency needs to be ex-
tended to models and analyses of within-person performance consis-
tency. Kane’s (1982) work on performance distributions is an isolated at-
tempt to understand the influence of ability and effort on individual per-
formance measurement and resultant organizational decision-making.
Individual consistency research would also provide information about
the frequency and duration of transition and maintenance stages that
workers experience. Murphy (1989) assumes that most jobs possess mul-
tiple transition stages that “occur at different times and last for different
durations” for different people (p. 195). Because progression through
these stages varies across individuals as well as jobs, it is important on
both theoretical and practical grounds that we are able to identifyworker
characteristics that might aid in the prediction of transition versus stable
stages of performance (Murphy, 1989).
Third, we agree with Murphy that consistency research should focus
on the job environment rather than the job itself. At the organizational
742 PERSONNEL PSYCHOLOGY
REFERENCES
Barrett GV, Caldwell MS, Alexander, RA. (1989). The predictive stability of ability re-
quirements for task performance: A critical reanalysis. Human Performance, 2(3),
167-181.
Bass, BM. (1962). Further evidence on the dynamic character of criteria. PERSONNEL
PSYCHOLOGY, 15,93-97.
Binning JE Barrett GV. (1989). Validity of personnel decisions: A conceptual analysis of
the inferential and evidential bases. Journal ofApplied Psychology, 74,478-494.
Cronbach LJ,Snow RE. (1977). Aptitudes and insmtctional methods. New York Irvington.
Dunnette MD. (1963). A modified model for test validation and selection research. In
Dreher G, Sackett P (Eds.), Perspectives on Employee Staffing and Selection (pp.9-
15). Homewood, I L Richard D. Irwin.
Ghiselli EE. (1956). Dimensional problems of criteria. Journal of Applied Psychology, 40,
14.
Ghiselli EE, Haire M. (1960). The validation of selection tests in light of the dynamic
nature of criteria. PERSONNEL PSYCHOLOGY, 13,225-231.
Guion RM. (1976). Recruiting, selection, and job placement. In Dunnette M (Ed.), Hand-
book of Industrial and Organizational Psychology (pp. 777-828). Chicago: Rand Mc-
Naliy.
Guion RM, Gibson WM. (1988). Personnel selection and placement. In Rosenzweig MR,
Porter LW (Eds.), Annual Review of Psychology, 39,349-74.
Helmreich RL, Sawin LL, Carsrud AL.(1986). The honeymoon effect in job performance:
Temporal increases in the predictive power of achievement motivation. Journal of
Applied Psychology, 71,185-188.
Henry RA, Hulin CL. (1987). Stability of skilled performance across time: Some general-
izations and limitations on utilities. Journal ofApplied Psychology, 72,457462.
Hollenbeck JR, Whitener EM. (1988). Reclaiming personality traits for personnel selec-
tion: Self-esteem as an illustrativecase. Journal of Management,14,81-91.
Hulin CL, Henry RA, Noon SL. (1990). Adding a dimension: Time as a factor in the
generalizability of predictive relationships. Psychological Bulletin, 107,328-340.
Hunter JE. (1983). Test validation for 12,000jobs: An application ofjob classification and
validiy generalization analysis to the GeneralApfitr.de Test Batrev. U.S.Employment
Service Test Research Report #45. Washington D C U.S. Department of Labor.
Ironson GH, Smith PC, Brannick MT, Gibson WM, Paul KB. (1989). Construction of a job
in general scale: A comparison of global, composite, and specific measures. Journal
of Applied Psychology, 74,193-200.
James, LR. (1973). Criterion models and construct validity for criteria. Psychological
Bulletin, 80,75-83.
Kane JS. (1982, November). Rethinking the problem of measuring performance: Some
new conclusions and a new appraisal method to fit them. Paper presented at the
Fourth Johns Hopkins University National Symposium on Educational Research,
Washington, DC.
Kanfer R, Ackerman PL. (1989). Motivation and cognitive abilities: An integrative/
aptitude-treatment interaction approach to skill acqu on.JournalofApplied Psy-
chology, 74,657-690.
McDaniel MA, Schmidt FL, Hunter JE. (1988). Job experience correlates of performance.
Journal of Applied Psychology, 73,327-330.
Murphy KR. (1989). Is the relationship between cognitive ability and job performance
stable over time? Human Performance, 2,183-200.
Prien EF! (1966). Dynamic character of criteria: Organizationalchange. JournalofAppZied
Psychology, 50,501-504.
Rambo WW, Chomiak AM, Price JM. (1983). Consistency of performance under stable
conditions of work. Journal of Applied Psychology, 68,78-87.
744 PERSONNEL PSYCHOLOGY
Rambo W,Chomiak AM, Rountree RI. (1987). Temporal intervals and the estimation of
the reliability of work performance data. Perceptual and Motor Skills, 64,791-798.
Ronan WN, Prien El? (1966). Towards a crirerion theov: A review and analysis of research
and opinion.Greensboro, NC: The Richardson Foundation.
Rothe HE (1978). Output rates among industrial employees. Journal ofApplied Psychol-
ogy, 63,4W6.
Schmidt FX,Hunter JE, Outerbridge AN. (1986). The impact of job experience and
ability on job knowledge, work sample performance and supervisory ratings of job
performance. Journal of Applied Psychology, 71,432-439.
Smith P.(1976). Behaviors, results, and organization effectiveness: The problem of crite-
ria. In Dunnette M (Ed.), Handbook of industrial and organizationalpsychology (pp.
745-775). Chicago: Rand McNally.
Wernimont PF, Campbell JF! (1968). Signs, samples, and criteria. Journal of Applkd
psycho lo^, 52,372-376.