You are on page 1of 6

Journal of Applied Psychology Copyright 1997 by the American Psychological Association. Inc.

1997, Vol. 82, No. 2, 247-252 0021-90HV97/S3.00

RESEARCH REPORTS

Overall Job Satisfaction: How Good Are Single-Item Measures?

John P. Wanous and Arnon E. Reichers Michael J. Hudy


Fisher College of Business, Ohio State University Ohio State University

A meta-analysis of single-item measures of overall job satisfaction (28 correlations from


17 studies with 7,682 people) found an average unconnected correlation of .63 (SO —
.09) with scale measures of overall job satisfaction. The overall mean correlation (cor-
rected only for reliability) is .67 (SD = .08), and it is moderated by the type of measure-
ment scale used. The mean corrected correlation for the best group of scale measures (8
correlations, 1,735 people) is .72 (5D = .05). The correction for attenuation formula
was used to estimate the minimum level of reliability for a single-item measure. These
estimates range from .45 to .69, depending on the assumptions made.

In general, single-item measures can be divided into single item may also be acceptable. Overall job satisfac-
two categories: (a) those measuring self-reported facts, tion appears to be an example, particularly because it has
such as years of education, age, number of previous jobs, a long history of single-item measures, beginning with
and so on, and (b) those measuring psychological con- the Faces Scale over 40 years ago (Kunin, 1955).
structs, such as job satisfaction. Measuring the former Scarpello and Campbell (1983) concluded that a sin-
with a single item is commonly accepted practice. How- gle-item measure of overall job satisfaction was prefera-
ever, the use of single-item measures for psychological ble to a scale that is based on a sum of specific job facet
constructs is typically discouraged, primarily because satisfactions. In the years since that article was published,
they are presumed to have unacceptably low reliability. however, the legitimacy of single-item measures has not
In fact, the use of single-item measures in academic re- increased. There are probably several reasons for this.
search is often considered a "fatal error" hi the review First, it is frequently said that one cannot estimate the
process. internal consistency reliability of single-item measures,
There are exceptions to the norm of using only scales and this alone is sometimes believed to be a sufficient
to measure psychological constructs, however. If the con- reason to limit or avoid their use. A second reason may
struct being measured is sufficiently narrow or is unam- be the increased use of structural equation models that
biguous to the respondent, a single-item measure may require estimation of the reliable variance.
suffice, as pointed out by Sackett and Larson (1990). for A third reason is that readers of Scarpello and Campbell
example, most expectancy theory researchers use a single (1983) may not have been convinced that the moderate
item to measure the perceived probability that effort leads correlations they reported between the single-item and
to performance (e.g., Ilgen, Nebeker, & Pritchard, 1981). a scale of job facet satisfactions were due to construct
For more complex psychological constructs (e.g., per- definitions, as they argued. Rather, readers may have con-
sonality), it is usually recommended that scales with mul- cluded that low reliability for the single-item measures
tiple items be used. However, there is a middle ground was a better explanation.
between the extremes of simple (expectancy) and com-
plex (personality) constructs where measurement with a
The Present Study

The present study is a meta-analysis (Hunter &


John P. Wanous and Arnon E. Reichers, Department of Man-
Schmidt, 1990) of research in which single-item measures
agement and Human Resources, Fisher College of Business,
of overall job satisfaction are correlated with scales mea-
Ohio State University; Michael J. Hudy, Department of Psychol-
suring overall job satisfaction. This was done for two
ogy, Ohio State University.
Correspondence concerning this article should be addressed reasons. First, the average correlation is more representa-
to John P. Wanous, Fisher College of Business, Ohio State Uni- tive of the population correlation than a correlation that
versity, Columbus, Ohio 43210. Electronic mail may be sent via is based on a single sample. Second, two pieces of infor-
Internet to wanous.l@osu.edu. mation from the meta-analysis are used to estimate the

247
248 RESEARCH REPORTS

reliability of single-item measures of overall job satisfac- two published studies (Ferratt, 1981; Kalleberg, 1974) because
tion by using a well-known formula (i.e., the correction their multiple-item measures were judged to be poor measures
for attenuation due to unreliability) in a new way. Spe- of overall job satisfaction. This is because an indirect measure
of job satisfaction was used ("should be-is now"), which
cifically, the average observed correlation between the sin-
involves calculating a raw discrepancy score. This measure of
gle-item and the scale measures of job satisfaction can
job satisfaction has been severely criticized (Wall & Payne,
be put into this formula, along with the estimated average
1973). Although judgment calls about the inclusion or exclusion
reliability of the scales used to measure overall job satis-
of data can affect a meta-analysis (Wanous, Sullivan, & Mali-
faction. By adding a value for the assumed underlying nak, 1989), omitting these two studies had no effect here, on
construct correlation between these two measures of over- the basis of a preliminary analysis.
all job satisfaction, it is possible to solve the equation for
its missing value, that is, the minimum reliability of the
Recorded Information
single-item measure. This procedure is described below.
(See also Wanous & Reichers, 1996b.) The sample size, type of job satisfaction measure, reliability
The well-known formula for the correction for attenua- of the scale measure, and the correlation between the single
tion can be found in most textbooks on psychometrics item and the scale were recorded for each study. The type of
(e.g., Nunnally & Bernstein, 1994, p. 257): single-item measure was subdivided into (a) 15 correlations
that used the Faces Scale, and (b) 13 correlations that used
some other type of single-item measure. The scales used to
measure overall job satisfaction were subdivided into (a) 8 cor-
relations that used a sum of items with a focus on overall job
satisfaction, (b) 9 correlations that used a sum of facets concern-
where rxy = correlation between variables x and y, r^ =
ing only the work itself, and (c) 11 correlations that used a sum
reliability of variable x, rw = reliability of variable y, and
of job facet satisfactions.
f^ = the assumed true underlying correlation between x
and y if both were measured perfectly.
This formula is most commonly applied to situations Meta-Analysis Calculations
where variables x and y are different constructs, for exam-
A software package developed by Switzer (1991) was used
ple, job satisfaction and job performance. However, the to make the meta-analytic calculations. This program uses the
formula can also be applied when both measured variables Hunter and Schmidt (1990) interactive procedure to produce
represent the same construct (i.e., job satisfaction, as done both observed and corrected means and standard deviations for
here). When this occurs, Nunnally (1978) reported the a set of correlations. The program also uses a bootstrap resam-
following: pling method to re-estimate the parameter estimates of the mean
and variance for comparison to the Hunter and Schmidt (1990)
The correlation between two such tests would be expected estimates. Furthermore, the bootstrap method yields 95% confi-
to equal the product of the terms in the denominator and dence intervals for each of these estimates (see Switzer, Paese, &
consequently rw would equal 1.00. . . . If rw were 1.00,
Drasgow, 1992, for a discussion of the bootstrap method).
rxy would be limited only by the reliabilities of the two
The specific corrections made here include (a) sampling error,
tests: rv = Vr^ • VJv (p. 220)
and (b) criterion reliability, in this case the reliability of the
scales, which averaged .88 for the six studies reporting this
Method information.

Literature Search
Results
The following search procedures were used. First, a compu-
terized search of the American Psychological Association's The Appendix shows the results of the literature search.
PsycLit database was conducted. Second, a manual search of Studies are listed alphabetically. In all, there were 17
studies cited in various published empirical articles and an un- different samples from 16 sources that reported a total of
published master's thesis (Thomas, 1993) on this topic was 28 correlations. There are two ways to report the sample
conducted. Third, a second computerized search was done with size: (a) the number of distinct persons in the 17 samples
the Business Abstracts database.
(N = 4,568; double counting of people is eliminated be-
cause some studies report multiple correlations), and ( b )
Criteria for Inclusion the number of persons for all 28 correlations (N = 7,682,
which includes double counting).
Studies selected for this meta-analysis reported correlations
between two measures of overall job satisfaction, one of which Table 1 shows the results of the meta-analysis. The
was a single-item measure. The second measure could be either mean observed correlation is .63 (SD = .09). When this
a sum of job facet satisfactions or the sum of several items is corrected for unreliability, the correlation is .67 (SD =
focused on overall job satisfaction. This led to the exclusion of .08). Total artifactual between-studies variance was 24%
RESEARCH REPORTS 249

Table 1
Analysis Results for Correlations Between Two Measures of Overall Job Satisfaction: Single-Item Measure and Scales

Bootstrap estimates
and confidence intervals (Cls)

Corrected mean r Corrected SD


Observed
Distribution N No. r, mean r SD mean r SD Estimate 95% CI Estimate 95% Q

All correlations 7,682 28 .63 .09 .67 .08 .67 .64-.70 .08 .04-. 11
All studies 4,568 17 .62 .08 .66 .08 .65 .61-.69 .07 .03-. 11
Type of single item
Measure
Faces 3,327 15 .62 .08 .66 .06 .67 .63 -.71 .06 .03-.09
Others 4,355 13 .64 .09 .68 .09 .68 .62-.72 .08 .00-. 14
Type of scale measure
Sum of facets 3,690 11 .64 .10 .68 .10 .68 .60- .72 .09 .00-.15
Work facet 2,257 9 .59 .06 .63 .05 .63 .59- .67 .04 .00-.07
Overall 1,735 8 .67 .06 .72 .05 .72 .67-.77 .04 .00-.07

of the variance, of which 18% was attributed to sampling relations is relatively small. However, in the present study,
error. the two methods yielded almost identical results.
As a check on the possible effect of double-counting Finally, it is possible to estimate the minimum reliabil-
people when multiple correlations are reported, the calcu- ity of a single-item measure using the correction for atten-
lations were recalculated on the 17 distinct samples. For uation formula, as described above. Several different esti-
those studies reporting more than one correlation, the mates were made, but in all cases the average observed
results were averaged and weighted by the average sample correlations were used, not the corrected ones, because
size. The results are quite similar to the results using all that would have resulted in double-correction and inflated
28 correlations. The corrected mean correlation is .66 (SD estimates.
= .08) for the 17 different samples. A total of eight estimates of the minimum reliability
Because 76% of the between-studies variance could not of a single-item measure are shown in Table 2. These are
be accounted for by measurable artifacts, a search for based on (a) which observed mean correlation was used,
moderators was undertaken. Two moderators were consid- (b) which assumption about the underlying construct cor-
ered: (a) how the single item was measured: Faces vs. relation was used, and (c) which scale reliability was
another type of single-item scale and (b) how the scales used. First, the mean observed correlation based on all
were measured: a sum of job facet satisfactions, a sum 28 that were found in our review (.63) was used, as is
of items focused on overall satisfaction, and a sum of the mean correlation for the subgroup using the best scale
facets concerning only the work itself, primarily the Work measure of overall job satisfaction, that is, those scales
scale of the JDI. using items that have a focus on overall satisfaction (. 67 ).
The type of single-item measure does not moderate the Second, one might assume that the true underlying corre-
overall correlation. However, the type of scale used to lation between these two measures of overall job satisfac-
measure overall job satisfaction does. Specifically, the dif- tion is 1.00. However, Scarpello and Campbell (1983)
ference between using a scale focused only on the work
itself versus a scale that is based on items focused on
overall satisfaction has the classical characteristics of a Table 2
moderator. That is, there is a mean difference between Estimates of Minimum Reliability
these two groups (r - .63 vs. .72) and the standard devia- Average observed correlation
tions for each group are smaller than when they are com- Assumptions about
bined. The third scale measure, sums of job facet satisfac- the construct correlation All studies Best scale
and about scale reliability (r = .63) (r = .67)
tions, had properties similar to the overall results.
Table 1 also shows the bootstrap re-estimates of the Assume true r = 1.00
mean and standard deviation from the meta-analysis and Reliability = .88 (observed) .45 .51
Reliability = .80 .50 .56
the 95% confidence intervals for each of these re-esti- Assume true r = .90
mates. This procedure was developed (Switzer; Pease, & Reliability = .88 (observed) .56 .63
Drasgow, 1992) to deal with the uncertainty associated Reliability = .80 .61 .69
M .63 .67
with such estimates, particularly when the number of cor-
250 RESEARCH REPORTS

challenged this assumption. As a result, estimates based reported thus far should be interpreted as questioning the
on a construct correlation of .90 are also shown. Third, use of well-constructed scales in comparison to a single-
only 6 of 17 studies reported scale reliability information, item measure. It should be interpreted, however, as a case
and it averaged .88, which is fairly high. Because this for the acceptability of single-item measures when either
average may not be representative, an alternative average the research question implies their use or when situational
reliability of .80 was also used. constraints limit or prevent the use of scales.
The estimates shown in Table 2 range from a low of One example of a research question suggesting the use
.45 to a high of .69, with an average of .57. Remember, of a single-item measure is the measurement of change
these estimates are of the minimum reliability for a single in overall job satisfaction. Suppose that a single question
item. The actual reliability could be higher, but it cannot is asked of respondents: ' 'How much did your overall job
be lower nor can it be estimated with any greater precision satisfaction change in the last year?" Such a single-item
than a minimum value. The effect of each of the three measure might be preferable to a scale, as the following
assumptions on the estimated minimum values approxi- example illustrates. Imagine a situation where Person A
mately equal, as shown in Table 2. has an average scale score of 3 by answering 3 to each
of 5 questions. Person B also has an average scale score
of 3 but answered two questions with 1, two more with
Discussion
5, and the last with 3. These differences are ignored if a
The observed correlations between single items and scale score is used.
scales averaged .63, and the corrected mean correlation There may also be practical limitations favoring the
of .67 further indicated convergent validity. This average use of a single-item measure. For example, if space on a
is much higher than the results reported by Scarpello and questionnaire is limited, researchers have to make choices
Campbell (1983) and seems to bolster their argument about how many items will be included. Specifically, the
that a single-item measure of overall job satisfaction is Minnesota Satisfaction Questionnaire (Short Form) has
acceptable. 20 job-facet items. If this popular scale is used only to
The average correlation is highest for the scales com- measure overall job satisfaction, 19 extra questions must
posed of items with an overall job satisfaction focus, as be answered.
compared with a sum of facets or the Work scale from There may be shorter alternatives to using a long list
the JDI. This is what one would expect to see because of job facets, such as scales that are focused on overall job
the focus of the items is directly on overall job satisfac- satisfaction (Brayfield & Rothe, 1951; Hoppock, 1935;
tion, and thus most similar to the wording of the single- Taylor & Bowers, 1972). However, these are all quite
item measures. wordy and consume precious space on a survey. The short-
The minimum level of reliability for single-item job est of these is probably the five-item General Satisfaction
satisfaction measures was also estimated by using the Scale from the Job Diagnostic Survey (JDS; Hackman &
standard correction for attenuation formula in a somewhat Oldham, 1975).
novel way. Reasonable estimates of this minimum value There might also be relevant cost considerations that
are those based on assuming that the true underlying con- would lead researchers to prefer a single-item measure.
struct correlation is .90 rather than 1.00, combined with For example, our university telephone poll charges $60
using the best measures available. These estimated mini- per item per sample (faculty, staff, students). Therefore,
mum reliabilities are .63 and .69, depending on the as- using the shortest of the scales (five items from the JDS)
sumption about the reliability of the scales. In fact, assum- would add $240 to the cost of measuring overall job satis-
ing the reliability is closer to .80 than .88 is also realistic faction for one sample and $720 for all three.
because an average reliability of .82 has been reported Finally, there may be issues of face validity. In particu-
for some job satisfaction scales (Fried & Ferris, 1987). lar, respondents may resent being asked questions that
Thus, a minimum estimated reliability for the single-item appear to be repetitious. This problem would be worse in
measure close to .70 is reasonable. an organization with poor employee relations. From a
It is also interesting to note that differences among management perspective, a single item is usually easier
single-item measures had no effect on the meta-analysis to understand than a scale score, which might appear as
results. However, differences in the ways that scales were academic nit-picking.
measured did make a difference in the results. Because In summary, if neither the research question nor the
of this difference, it seems reasonable to conclude that research situation suggest the use of a single-item job
the single-item measures are more robust than the scale satisfaction measure, then choosing a well constructed
measures of overall job satisfaction. scale makes sense. However, if the use of a single item is
Despite the above data and arguments, there are still indicated, researchers may do so in the knowledge that
good reasons for preferring scales to single items. Nothing they can be acceptable. The use of single-item measures
RESEARCH REPORTS 251

should not be considered fatal flaws in the review process. W. R., & Feild, H. S. (1988). Job performance and turnover
Rather, their appropriateness for a particular piece of re- decisions: Two field studies. Journal of Management, 14,
403-414.
search should be evaluated. Future research should be
Nunnally, J. C. (1978). Psychometric theory (2nd ed.). New
directed toward examining the validity and reliability of
%rk: McGraw-Hill.
other single-item measures.
Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric theory
(3rd ed.). New %rk: McGraw-Hill.
References *0'Reilly, C. A., & Roberts, K. H. (1973). Job satisfaction
among Whites and non-Whites: A cross-cultural approach.
References marked with an asterisk indicate studies included in Journal of Applied Psychology, 57, 295-299.
the meta-analysis. *O'Reilly, C. A., & Roberts, K. H. (1978). Supervisor influence
and subordinate mobility aspirations as moderators of consid-
*Adler, S., & Golan, J. (1981). Lateness as a withdrawal behav-
eration and initiating structure. Journal of Applied Psychol-
ior. Journal of Applied Psychology, 66, 544-554.
ogy, 63, 96-102.
Brayfield, A. H., & Rothe, H. F. (1951). An index of job satis-
*Rosse, J. G., & Hulin, C. L. (1985). Adaptation to work: An
faction. Journal of Applied Psychology, 35, 307-311.
analysis of employee health, withdrawal, and change. Organi-
*Brief, A. P., & Robereon, L. (1989). Job attitude organization:
zational Behavior and Human Decision Processes, 36, 324-
An exploratory study. Journal of Applied Social Psychology,
347.
19, 717-727. *Russell, J. C., & Farrar, W. E. (1978). Job satisfaction in uni-
*Dunham, R. B., & Herman, J. B. (1975). Development of a versity employees: A field test of attitude theory. Psychologi-
female faces scale for measuring job satisfaction. Journal of cal Reports, 42, 1247-1254.
Applied Psychology, 60, 629-631. Sackett, P. R., & Larson, J. R., Jr. (1990). Research strategies
Ferratt, T. W. (1981). Overall job satisfaction: Is it a linear and tactics in industrial and organizational psychology. In
function of facet satisfaction? Human Relations, 34, 463- M. D. Dunnette & L. M. Hough (Eds.), Handbook of indus-
473. trial and organizational psychology (2nd ed., Vol. 1, pp. 419-
Fried, Y., & Ferris, G. R. (1987). The validity of the job charac- 489). Palo Alto, CA: Consulting Psychologists Press.
teristics model: A review and meta-analysis. Personnel Psy- *Scarpello, V., & Campbell, J. P. (1983). Job satisfaction: Are
chology, 40, 287-322. all the parts there? Personnel Psychology, 36, 577-600.
Hackman, J. R., & Oldham, G. R. (1975). Development of the *Smith, P. C., Kendall, L. M., & Hulin, C. L. (1969). The mea-
Job Diagnostic Survey. Journal of Applied Psychology, 60, surement of satisfaction in work and retirement. Chicago:
159-170. Rand McNally.
*Hom, P. W., Katerberg, R., & Hulin, C. L. (1979). Compara- Switzer, F. (1991). VGBOOT2.0. Department of Psychology,
tive examination of three approaches to the prediction of turn- Clemson University, Clemson, South Carolina.
over. Journal of Applied Psychology, 64, 280-290. Switzer, F. S., DI, Pease, P. W., & Drasgow, F. (1992). Bootstrap
Hoppock, R. (1935). Job satisfaction. New \brk: Harper. estimates of standard errors in validity generalization. Journal
Hunter, J. E., & Schmidt, F. L. (1990). Methods of meta-analy- of Applied Psychology, 77, 123-129.
sis. Newbury Park, CA: Sage. Taylor, J. C., & Bowers, D. G. (1972). Survey of organizations.
Ilgen, D. R., Nebeker, D. M., & Pritchard, R. D. (1981). Expec- Ann Arbor: Institute for Social Research, University of
tancy theory measures: An empirical comparison in an experi- Michigan.
. mental simulation. Organizational Behavior and Human Per- Thomas, A. L. (1993). Job satisfaction and its consequences:
formance, 28, 189-223. A meta-analytic approach. Unpublished master's thesis, Ohio
*Ironson, G. H., Smith, P. C., Brannick, M. T., Gibson, State University, Department of Psychology, Columbus.
W. M., & Paul, K. B. (1989). Construction of a job in general Wall, T. D., & Payne, R. (1973). Are deficiency scores defi-
scale: A comparison of global, composite, and specific mea- cient? Journal of Applied Psychology, 58, 322-326.
sures. Journal of Applied Psychology, 74, 193-200. *Wanous, J. P., & Lawler, E. E. III. (1972). Measurement and
Kalleberg, A. L. (1974). A causal approach to the measurement meaning of job satisfaction. Journal of Applied Psychology,
of job satisfaction. Social Science Research, 3, 299-322. 56, 95-105.
Kunin, T. (1955). The construction of a new type of attitude *Wanous, J. P., & Reichers, A. E. (1996a). [Correlations be-
measure. Personnel Psychology, 8, 65-78. tween a single-item measure of overall job satisfaction and
*LaRocco, J. M. (1983). Job attitudes, intentions, and turnover: the sum of job facet satisfactions.] Unpublished raw data.
An analysis of effects using latent variables. Human Rela- Wanous, J. P., & Reichers, A. E. (1996b). Estimating the relia-
tions, 36, 813-825. bility of a single-item measure. Psychological Reports, 78,
*Miller, H. E., Katerberg, R., & Hulin, C. L. (1979). Evaluation 631-634.
of the Mobley, Homer, and Hollingsworth model of employee Wanous, J. P., Sullivan, S. E., & Malinak, J. (1989). The role
turnover. Journal of Applied Psychology, 64, 509-517. of judgment calls in meta-analysis. Journal of Applied Psy-
*Mossholder, K. W., Bedian, A. G., Morris, D. R., Giles, chology, 74, 259-264.

{Appendix follows on next page)


252 RESEARCH REPORTS

Appendix

Studies Used in the Mela-Analysis

Study Single item" Scale'

Adler & Golan (1981) 82 .98 Faces Sum of facets .74 (JDI)
Brief & Roberson (1989) 144 .92 Faces Sum of facets .71 (JDI)
144 .89 .71 (MSQ-SF)
Dunham & Herman (1975) 113 Faces Overall .75 (Hoppock Job Satisfaction Blank)
Horn et al. (1979) 373 Faces Work facet .53 (JDI)
Ironson et al. (1989) 227 Faces Overall .75 (JIG) .65 (B-R) .68 (JDI)
227 Item (A) Overall .76 (JIG) .67 (B-R) .68 (JDI)
227 Item Work facet .67 (JIG) .60 (B-R) .59 (JDI)
LaRocco (1983) 260 .92 Item Overall .58 (JDS)
Miller et al. (1979) 235 .88 Faces Work facet .63 (JDI)
225 Faces Work facet .59 (JDI)
Mossholder et al. (1988) 220 .89 Item Sum of facets .71 (MSQ-SF)
O'Reilly & Roberts (1973) 139 Faces Work facet .45 (JDI)
O'Reilly & Roberts (1978) 562 Faces Work facet .59 (JDI)
Rosse & Hulin (1985) 42 .81 Faces Work facet .44 (JDI)
Russell & Farrar (1978) 507 .91 Faces .63 (JDI)
Scarpello & Campbell (1983) 185 Item Sum of facets .32 (MSQ)
185 Item (Yes-No) .42 (MSQ)
Smith et al. (1969) 80 Faces .54 (JDI)
Wanous & Lawler (1972) 208 Sum of facets .60
Wanous & Reichers (1996a) 969 .86 Item Sum of facets .68
966 .77 .68

a
Single item measures include (a) the Faces scale—Faces, (b) a numerical item—Item, (c) an adjective item—Item (A), and (d) an item with a
yes-no response—Item (Yes-No). b Scale measures include (a) a sum of facet satisfaction items—Sum of facets, (b) a sum of overall-focused
items—Overall, and (c) a sum of facet satisfaction items concerning only the work itself—Work facet. c JDS = Job Diagnostic Survey; JDI =
Job Descriptive Index; MSQ = Minnesota Satisfaction Questionnaire; MSQ-SF = Short Form of the MSQ; B-R = Brayfield-Rothe Scale.

Received May 23, 1996


Revision received September 13, 1996
Accepted September 17, 1996 •

You might also like