You are on page 1of 16

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/220260414

Measuring Information Systems Service Quality: Concerns


on the Use of the SERVQUAL Questionnaire

Article  in  MIS Quarterly · June 1997


DOI: 10.2307/249419 · Source: DBLP

CITATIONS READS
484 1,053

3 authors, including:

Leon A. Kappelman V.R. Prybutok


University of North Texas University of North Texas
84 PUBLICATIONS   2,406 CITATIONS    241 PUBLICATIONS   6,390 CITATIONS   

SEE PROFILE SEE PROFILE

Some of the authors of this publication are also working on these related projects:

A Trust Model of Clinic Appointment Systems View project

Analysis of Late Response Bias Structural Equation Modeling View project

All content following this page was uploaded by Leon A. Kappelman on 17 June 2014.

The user has requested enhancement of the downloaded file.


SERVOUAL Concems

Measuring Abstract

Information Systems A recent MIS Quarterly article rightfully points


out that service is an important part of the role
Service Quaiity: of the information systems (IS) department and
that most IS assessment measures have a
Concerns on the Use product orientation (Pitt et al. 1995). The article
went on to suggest the use of an IS-context-
of the SERVQUAL modified version of the SERVQUAL instrument
to assess the quality of the services supplied by
Questionnaire^ an information services provider (Parasuraman
et al. 1985, 1988, 1991).^ However, a number
of problems with the SERVQUAL instrument
have been discussed in the literature (e.g.,
Babakus and Boiler 1992; Carman 1990;
By: Thomas P. Van Dyke Cronin and Taylor 1992, 1994; Teas 1993).
Department of Information Systems This article reviews that literature and discusses
some of the implications for measuring sen/ice
and Technologies
quality In the information systems context.
College of Business and Economics Findings indicate that SERVQUAL suffers from
Weber State University a number of conceptual and empirical difficul-
3804 University Circie ties. Conceptual difficulties include the opera-
Ogden, Utah 84408-3804 tionalization of perceived service quality as a
difference or gap score, the ambiguity of the
U.S.A.
expectations construct, and the unsuitability of
tvandyke @ weber.edu using a single measure of service quality across
different industries. Empirical problems, which
may be linked to the use of difference scores,
include reduced reliability, poor convergent
Leon A. Kappelman validity, and poor predictive validity. This sug-
Business Computer Information gests that (1) some alternative to difference
Systems Department scores is preferable and should be utilized; (2) if
Coilege of Business Administration used, caution should be exercised in the inter-
pretation of iS-SERVQUAL difference scores;
University of North Texas
and (3) further work is needed in the develop-
Denton, Texas 76203-3677 ment of measures for assessing the quality of
U.S.A. iS services..
kapp@unt.edu

Keywords: IS management, evaluation, mea-


surement, service quality, user attitudes,
Victor R. Prybutok user expectations
Business Computer information
Systems Department
iSRL Categories: AI0104, AlOlOg, A104,
Coilege of Business Administration E10206.03, GB02, GB07
University of North Texas
Denton, Texas 76203-3677 The terms "information services provider" or "IS services
U.S.A. provider* are used to refer to any provider of information
systems services. This inciudes the information systems
function within an organization as weii as extemai vendors
'Robert Zmud was the accepting senior editor for this paper. of information systems products and/or services.

MIS Ouarterly/June 1997 195


Research Note

Introduction SERVOUAL instrument in general and exam-


ines the Implications of these difficulties to the
Due to the growth of outsourcing, end-user- use of the instrument in an IS context.
controlled information assets, joint ventures,
and other alternative mechanisms by which
organizations are meeting their need for infor-
mation systems services, IS managers are The SERVQUAL Instrument:
increasingly concerned about improving the Problems Identified in the
perceived (as well as actual) service quality of
the IS function (Kettinger and Lee 1994). In Literature
recent years, the use of SERVOUAL-based
instruments has become increasingly popular The difficulties with the SERVOUAL instrument
with information systems (IS) researchers. identified in the literature can be grouped into
However, a review of the literature suggests two main categories: (1) conceptual and (2)
that the use of such instruments may result in empirical; although, the boundary between
a number of measurement problems. them blurs because they are closely inter-relat-
ed. The conceptual problems center around
A recent article makes several important con- (1) the use of two separate instruments, one
tributions to the assessment and evaluation of for each of two constructs (i.e., perceptions
the effectiveness of information systems (IS) and expectations), to operationalize a third
departments in organizations (Pitt et al. 1995). conceptually distinct construct (i.e., perceived
The article: service quaiity) that is itself the result of a com-
plex psychological process; (2) the ambiguity
1. Points out that, although service is an of the expectations construct; and (3) the suit-
important part of the role of the IS depart- ability of using a single instrument to measure
ment, most IS assessment measures have service quality across different industries (i.e.,
a product orientation. content validity). The empirical problems are,
by and large, the result of these conceptual dif-
ficulties, most notably the use of difference
2 Proposes an extension of the categorization
scores, in conjunction with the a theoretical
of IS success measures (DeLone and
nature of the process used in the construction
McLean 1992) to include service quality.
of the original five dimensions of service quali-
ty. The empirical difficulties most often attrib-
3. Proposes the use of the SERVOUAL instru- uted to the SERVOUAL instrument include low
ment from marketing (Parasuraman et al. reliability, unstable dimensionality, and poor
1985, 1988, 1991) to operationalize the IS convergent validity. A review of these concep-
service quality construct and modify the tual and empirical problems should serve to
wording of the instrument to better accom- caution those who wish to use SERVOUAL to
modate its use in an IS context. measure the service quality of an information
system provider.
4. Adapts and augments a theory regarding
the determinants of service quality expecta-
tions to an IS context and offers ideas for Conceptual difficuities with
future research.
SERVQUAL
A number of studies, however, identify poten-
tial difficulties with the SERVOUAL instrument Subtraction as a "Simuiation" of a
(e.g., Babakus and Boiler 1992; Carman 1990; Psychoiogicai Process
Cronin and Taylor 1992, 1994; Teas 1993,
1994). This research note reviews some of the Many of the difficulties associated with the
literature regarding difficulties with the SERVQUAL instrument stem from the opera-

196 MIS Quarteriy/June 1997


SERVQUAL Concems

tionalization of a service quality construct that hopes to receive (e.g., Parasuraman et al.
is theoretically grounded in a discrepancy or 1985,1988, 1991; ZeithamI et al. 1993). These
gap model. In conceptualizing service quality, multipie definitions and corresponding opera-
Parasuraman et al. 1985, 1988, 1991, 1994b) tionalizations of "expectations" in the
use the "service quality model," which posits SERVQUAL literature result in a concept that
that one's perception of service quality is the is loosely defined and open to multiple inter-
result of an evaluation process whereby "the pretations (Teas). Yet even when concise defi-
customer compares . . . the perceived service nitions are provided, various interpretations of
against the expected semce" (Gronroos 1984, the expectations construct can result in poten-
p. 37). tially serious measurement validity problems.

Rather than develop an instrument to directly The conceptualization of "expectations" con-


measure the perception of service quaiity that sistent with the SERVQUAL model is the vec-
is the outcome of this cognitive evaluation tor attribute interpretation—"that is one on
process, the SERVQUAL instrument which a customer's ideal point is at an infinite
(Parasuraman et al. 1988, 1991) separately level" (Parasuraman et al. 1994, p. 116).
measures the expected level of service and Unfortunately, as the proportion of extreme
the experienced level of service. Then service responses (e.g., seven on a seven point scale)
quality scores are calculated as the difference increases, the expectation scores become less
between these two measures. These three useful as an increasing proportion of the varia-
sets of scores are commonly referred to as tion in difference-based SERVQUAL scores is
expectation (E), perception (P), and due only to changes in perception scores.
SERVQUAL (whereby, SERVQUAL = P - E).
Although not without precedent, and certainly Teas (1993) found three different interpretations
worthy of fair empirical evaluation, the implicit of "expectations" derived from an analysis of
assumption that subtraction accurately por- follow-up questions to an administration of the
trays this cognitive process seems overly sim- SERVQUAL questionnaire. Qne interpretation
plistic. Even if one fully accepts this discrepan- of expectations is as a forecast or prediction.
cy model of experiences vis-^-vis expectations The forecast interpretation of expectations can-
as indicative of the general process whereby not be discriminated from the disconfirmed
one arrives at an evaluation of a service expe- expectations model of consumer satisfaction
rience, the notion that the specific mechanism (Oliver 1980). This interpretation is inconsistent
is mereiy subtraction does not logically follow. with the definition of service quality put forth by
The use of differences was, and remains, an Parasaraman et al. (1988) and results in a dis-
operational decision. Regrettably, it does not criminant validity problem with respect to con-
appear to have been a particularly good one. sumer satisfaction. A second interpretation of
The direct measurement of one's perception of expectations is as a measure of attribute impor-
service quality that is the outcome of this cog- tance. When respondents use this interpreta-
nitive evaluation process seems more likely to tion, the resulting perception-minus-expectation
yield a valid and reliable outcome. If the dis- scores exhibit an inverse relationship between
crepancy is what one wants to measure, then attribute importance and perceived service
one should measure it directly. quality, all other things being equal.

The third interpretation identified by Teas


(1993) is the "classic ideal point" concept.
Ambiguity of the "Expectations" Construct Parasuraman et al. (1994) describe this when
they note that "the P-E [i.e., perceptions-
Teas (1994) notes that SERVQUAL expecta- minus-expectations] specification could be
tions are variously defined as desires, wants, problematic when a service attribute is a das-
what a service provider should possess, nor- sic ideai point attribute—that is one on which a
mative expectations, ideal standards, desired customer's ideal point is at a finite level and
service, and the level of service a customer therefore, performance beyond which will dis-

MIS Quarterly/June 1997 197


Research Note

please the customer (e.g., friendiiness of a cate that perceptions are influenced by both
salesperson in a retail store)" (p. 116). This will and should expectations, but in opposite
interpretation of expectations results in an directions. Increasing w///expectations leads to
inverse of the relationship between the a higher perception of service quality whereas
SERVQUAL score, calculated as perception an increasing expectation of what should be
minus expectation (P - E), and actual service delivered during a service encounter will actu-
quaiity for all values when perception scores ally decrease the ultimate perception of the
are greater than expectation scores (i.e., P > quality of the service provided (Boulding et al.
E). This interpretation is consistent with the 1993). Not only do these findings faii to sup-
finding that user satisfaction scores were high- port the gap model of service quality, but these
est when actual user participation was in con- results also demonstrate the wildly varying
gruence with the user's need for participation, impact of different interpretations of the expec-
rather than merely maximized (Doll and tations construct.
Torkzadeh 1989).
Different methods to operationalize "expecta-
These various interpretations of the "expecta- tions" in developing their IS versions of
tion" construct lead to a number of measure- SERVQUAL have been used (Pitt el al. 1995;
ment problems. The findings suggest that a Kettinger and Lee 1994). Qne study used the
considerable portion of the variance in the instructions to the survey to urge the respon-
SERVQUAL instrument is the result of mea- dents to "think about the kind of IS unit that
surement error induced by respondent's vary- would deliver excellent quality of service" (Pitt
ing interpretations of the "expectations" con- et al. 1995). The items then take a form such
struct (Teas 1993). as: El They wiii have up-to-date hardware and
software. Whereas the second study (Kettinger
Three separate types of expectations have and Lee 1994) used the form: E1 Excellent
been described (Boulding et al. 1993): (1) the college computing services will have up-to-
will expectation, what the customer believes date equipment
will happen in their next service encounter; (2)
the should expectation, what the customer Recall that some respondents to SERVQUAL
believes should happen in the next service were found to interpret expectations as fore-
encounter; and (3) an ideal expectation, what casts or predictions (Teas 1993). This interpre-
a customer wants in an ideal sense. The ideai tation corresponds closely with the will expec-
interpretation of expectation is often used in tation (Boulding et al. 1993). It is easy to see
the SERVQUAL literature (Boulding et al. how this interpretation might be formed espe-
1993). Boulding et al. (1993) differentiate cially with the "They will" phrasing (Pitt et al.
between shouid and ideai expectations by stat- 1995). Unfortunately, the impact of the will
ing that what customers think shouid happen expectation on perceptions of service quality is
may change as a result of what they have opposite from that intended by the
been told to expect by the service provider, as SERVQUAL authors and the (P-E) or gap
well as what the consumer views as reason- model of service quality (Boulding et al. 1993).
able and feasible based on what they have
been told and their experience with the firm or In summary, a review of the literature indicates
a competitor's service. In contrast, an ideal that respondents to SERVQUAL may have
expectation may "be unrelated to what is rea- numerous interpretations of the expectations
sonable/feasible and/or what the service construct and that these various interpretations
provider tells the customer to expect" have different and even opposite impacts on
(Boulding et al. 1993, p. 9). A series of experi- perceptions of service quality. Moreover, some
ments demonstrated results that were incom- of the findings demonstrate that expectations
patibie with the gap model of service quality influence only perceptions and that percep-
(Boulding et al. 1993). Instead, the results tions alone directly influence overall service
demonstrate that service quality is influenced quality (Boulding et al. 1993). These findings
only by perceptions. Moreover, the results indi- fail to support the (P-E) gap model of service

198 MIS Quarterly/June 1997


SERVOUAL Concems

quality and indicate that the use of the expec- findings suggest that neither the UIS nor
tations construct as operationalized by SERVQUAL alone can capture all of the fac-
SERVQUAL-based instruments is problematic. tors which contribute to perceived service
quality in the IS domain. For example, items
contained in the UIS include the degree of
training provided to users by the IS staff, the
Applicability of SERVQUAL Across level of communication between the users and
Industries the IS staff, and the time required for new sys-
tems development and implementation, all of
Another often mentioned conceptual problem which possess strong face validity as determi-
with SERVQUAL concerns the applicability of nants of IS service quality. In addition,
a single instrument for measuring service qual- Kettinger and Lee dropped the entire tangibies
ity across different industries. Several dimension from their IS version of SERVQUAL
researchers have articulated their concerns on based on the results of confirmatory factor
this issue. A study of SERVQUAL across four analysis. These finding contradict the belief
different industries found it necessary to add that all dimensions of SERVQUAL are relevant
as many as 13 additional items to the instru- and that there are of no unique features of the
ment in order to adequately capture the ser- IS domain not included in the standard
vice quality construct in various settings, while SERVQUAL instrument (Pitt et al. 1995). It is
at the same time dropping as many as 14 difficult to argue that items concerning the
items from the original instrument based on manner of dress of IS employees and the visu-
the results of factor analysis (Carman 1990). al attractiveness of IS facilities (i.e., tangibles)
The conclusion arrived at was that consider- should be retained as important factors in the
able customization was required to accommo- IS domain while issues such as training, com-
date differences in service settings. Another munication, and time to complete new systems
study attempted to utilize SERVQUAL in the development are excluded. We agree that
banking industry (Brown et al. 1993). The using a single measure of service quality
authors were struck by the omission of items across industries is not feasible (Dabholkar et
which they thought a priori would be critical to - al. 1996) and therefore future research should
subject's evaluation of service quality. They involve the development of industry-specific
concluded that it takes more than simple adap- measures of service quality.
tation of the SERVQUAL items to effectively
address service quality across diverse set-
tings. A study of sen/ice quality for the retail
sector also concluded that utilizing a single
Empirical difficulties with the
measure of service quality across industries is SERVQUAL instrument
not feasible (Dabholkar et al. 1996).
A difference score Is created by subtracting
Researchers of service quality in the informa- the measure of one construct from the mea-
tion systems context appear to lack consensus sure of another in an attempt to create a mea-
on this issue. Pitt et al. (1995) state that they sure of a third distinct construct. For example,
could not discern any unique features of IS in scoring the SERVQUAL instrument, an
that make the standard SERVQUAL dimen- expectation score is subtracted from a percep-
sions inappropriate nor could they discern any tion score to create such a "gap" measure of
dimensions with some meaning of service service quality. Even if one assumes that the
quality in the IS domain that had been exclud- discrepancy theory is correct and that these
ed from SERVQUAL. Kettinger and Lee are the only (or at least, the last) two inputs
(1994), however, found that SERVQUAL into this cognitive process, it still raises the
should be used as a supplement to the UIS question: Can calculated difference scores
(Baroudi and Qrlikowski 1988) because that operationalize the outcome of a cognitive dis-
instrument also contains items that are infipor- crepancy? It appears that several problems
tant determinants of IS service quality. Their with the use of difference scores make them a

MiS Ouarteriy/June 1997 199


Research Note

poor measure of psychological constructs This formula shows that as the correlation of
(e.g., Edwards 1995; Johns 1981; Lord 1958; the component scores increases, the reliability
Peter et al. 1993; Wall and Payne 1973). of the difference scores is decreased. An
Among the difficulties related to the use of dif- example was provided where the reliability of
ference measures discussed in the literature the difference score formed by subtracting one
are low reliability, unstable dimensionality, and component from another with an average relia-
poor predictive and convergent validities. bility of .70, and a correlation of .40, is only .50
(Johns 1981). Thus, while the average reliabili-
ty of the two components is .70, which is con-
sidered acceptable (Pitt et al. 1995; cf.,
Nunnally 1978), the correlation between the
Reiiabiiity Probiems With Difference Scores
components reduces the reliability of the differ-
Many studies demonstrate that Cronbach's ence score to a level that most researchers
would consider unacceptable (Peter et al.
(1951) alpha, a widely-used method of esti-
1993).
mating instrument reliability, is inappropriate
for difference scores (e.g., Cronbach and
An example of the overestimatlon for the relia-
Furby 1970; Edwards 1995; Johns 1981; Lord
bility caused by the misuse of Cronbach's
1958; Peter et al. 1993; Prakash and alpha can be found in the analysis of service
Lounsbury 1983; Wall and Payne 1973). This quality for a computer manufacturer
is because the reliability of a difference score (Parasuraman et al. 1994a; see Table 1). Note
is dependent on the reliability of the compo- that Cronbach's alpha consistently overesti-
nent scores and the correlation between them. mates the actual reliability for the difference
The correct formula for calculating the reliabili- scores of each dimension (column 2). Also
ty of a difference score (rD) is: note that the use of the correct formula for cal-
culating the reliability of a difference score has
demonstrated that the actual reliabilities for the
SERVQUAL dimensions may be as much as
.10 lower than reported by researchers incor-
rectly using Cronbach's alpha. In addition,
these findings show that the non-difference,
direct response method results in consistently
where r,, and r^ are the reliabilities of the two
higher reliability scores than the (P-E) differ-
component scores,CT,^and a^^ are the vari-
ence method of scoring.
ances of the component scores, and r,2 is the
correlation between the component scores These results have important implications for
(Johns 1981). the IS-SERVQUAL (Pitt et al. 1995).

Table 1. Reliability of SERVQUAL: The iUlisuse of Cronbach's Alpha

Johns' a for
A Priori Cronbachs' a Cronbachs' a Differences
Dimensions (Non-Difference) (Difference) (Difference)
Tangibles .83 .75 .65
Reliability .91 .87 .83
Responsiveness .87 .84 .81
Assurance .86 .81 .71
Empathy .90 .85 .81
Note: Difference scores calculated as perception minus expectation (P - E).

200 MIS Quarterly/June 1997


SERVOUAL Concems

Cronbach's alpha, which consistently overesti- values (ranging from .72 to .81) compared to
mates the reliability of difference scores, was the SERVOUAL difference scores (ranging
used incorrectly. Even when using the inflated from .51 to .71).
alpha scores, Pitt et al. note that two of three
reliability measures for the tangibles dimension The predictive validity of difference scores, a
fall below the 0.70 level required for commer- non-difference direct response score, and the
cial applications. Had they utilized the appro- perceptions only scores for SERVOUAL in the
priate modified alpha, they may have conclud- context of a financial institution hae been com-
ed that the tangibles dimension is not reliable pared (Brown et al. 1993). Correlation analysis
in the IS context, a finding which would have was performed between the various scores
been consistent with the results of Kettinger and a three-item behavioral intentions scale.
and Lee (1994). Behavioral intentions include such concepts as
whether the customer would recommend the
A review of the literature clearly indicates that financial institution to a friend or whether they
by utilizing Cronbach's alpha, researchers tend would consider the financial institution first
to overestimate the reliabilities of difference when seeking new services. The results of the
scores especially when the component scores study show that both the perceptions only (.31)
are highly correlated: Such is the case with the and direct response (.32) formats demonstrat-
SERVOUAL instrument (Peter et al. 1993). ed higher correlations with the behavioral
intentions scale than did the traditional differ-
ence score (.26).

Predictive and Convergent Vaiidity issues The superior predictive and convergent validity
of perception-only scores was confirmed
With Difference Scores
(Cronin and Taylor 1992). Those results indi-
Another problem with the SERVOUAL instru- cated higher adjusted r-squared values for per-
ment concerns the poor predictive and conver- ception-only scores across four different indus-
gent validities of the measure. Convergent tries. The perception component of the per-
validity Is concerned with the extent to which ception-minus-expectation score consistently
multiple measures of the same construct agree performs better as a predictor of overall ser-
with each other (Cambell and Fiske 1959). vice quality than the difference score itself
Predictive validity refers to the extent to which (Babakus and Boiler 1992; Boulding et al.
scores of one construct are empirically related 1993; Cronin and Taylor 1992; Parasuraman
to scores of other conceptually-related con- etal. 1991).
structs (Bagozzi et al. 1992; Kappelman 1995;
Parasuraman et al. 1991).

One study reported that perceptions-only Unstabie Dimensionaiity of the SERVQUAL


SERVOUAL scores had higher correlations Instrument
with an overall service quality measure (i.e.,
convergent measure) and with complaint reso- The unstable nature of the factor structure for
lution scores (i.e., the predictive measure) the SERVOUAL instrument may be related to
than did the perception-minus-expectation dif- the atheoretical process by which the original
ference scores used with SERVOUAL dimensions were defined. The SERVOUAL
(Babakus and Boiler 1992). A different study , questionnaire is based on a multi-dimensional
performed regression analyses in which an model (i.e., theory) of service quality. A 10
overall single-question service quality rating dimensional model of service quality based
was regressed separately on both difference on a review of the service quality literature
scores (i.e., perception minus expectation) and the extensive use of both executive and
and perception-only scores (Parasuraman et focus group interviews was developed
al. 1991). The perception-only SERVOUAL (Parasuraman et al. 1985). During instrument
scores produced higher adjusted r-squared development, Parasuraman et al. (1988)

MIS Quarterly/June 1997 201


Research Note

Table 2. Unstable Dimensionality of SERVQUAL

Factor
Study Instrument Analysis Structure
Carman (1990) Four modified Principal axis factor Five to nine factors
SERVOUALs using analysis with oblique
12-21 of the original rotation
items
Bresinger and Lambert Original 22 items Principal axis factor Four factors with
(1990) anaiysis with obiique eigenvalues > 1
rotation
Parasuraman, ZeithamI, Original 22 items Principal axis factor Five factors, but different
and Berry (1991) analysis with oblique from a priori model.
rotation Tangibles dimension spiit
into two factors, while respon-
siveness and assurance
dimensions loaded on a single
factor.
Finn and Lamb (1991) Originai 22 items LISREL confirmatory Five-factor model
factor analysis had poor fit.
Babakus and Boiler Original 22 items (1) Principal axis factor (1) Five-factor modei
(1991) analysis with oblique not supported
rotation.
(2) Confirmatory factor (2) Two factors
analysis
Cronin and Taylor (1992) Originai 22 items Principal axis factor Unidimensional
analysis with oblique structure
rotation
"Van Dyke and Popelka 19 of original 22 items Principal axis factor Unidimensional
(1993) analysis with oblique structure
rotation
*Pitt, Watson, and Kavan Original 22 items Principal components (1) Financial institution
(1995) and maximum iikelihood seven-factor model with
with verimax rotation tangibies and empathy spWt
into two.
(2) Consulting firm five-
factors, none matching the
original.
(3) Information systems
service firm—three-factor
model.
*Kettinger and Lee Original 22 items LISREL confirmatory Four-factor model, tangibles
(1994) factor analysis dimension dropped.
•Kettinger, Lee, and Lee Original 22 items Principal axis factor (1) Korea, three-factor
(1995) analysis with oblique model, tangibies retained.
rotation (2) Hong Kong, four-factor
model, tangibies retained.
"Measured information systems service quality.

202 MiS Quarterly/June 1997


SERVQUAL Concerns

began with 97 paired questions (i.e., one for SERVQUAL not match the proposed model, its
expectation and one for perception), items factor structure varied across settings.
(i.e., question pairs) were first dropped on the Analysis of the data from the consulting firm
basis of within-dimension Cronbach coeffi- resuited in a five-factor model although none
cient aiphas, reducing the pool to 54 question of these matched the originai a priori factors.
pairs. More items were then dropped or re-
The factor analysis of the information systems
assigned based on oblique-rotation factor
business data resulted in the extraction of only
loadings and within-dimension Cronbach
coefficient alphas resulting in a 34 paired-item three factors.
instrument with a proposed seven-dimension-
LISREL confirmatory factor analysis was used
al structure. A second data collection and
analysis with this "revised" definition and on SERVQUAL data collected from users (i.e.,
operationalization of service quaiity resulted students) of a college computing services
in the 22 paired-item SERVQUAL instrument department (Kettinger and Lee 1994).
with a proposed five-dimensional structure. Analysis of this data resuited in a four factor
Two of these five dimensions contained items solution. The entire tangibies dimension was
representing seven of the original 10 dimen- dropped. An IS version of SERVQUAL was
sions. We are cautioned, however, that those used in a cross-national study (Kettinger et ai.
who wish to interpret factors as real dimen-
1995). Results of exploratory common factor
sions shoulder a substantial burden of proof
analysis with oblique rotation indicated a
(Gronbach and Meehl 1955). Moreover, such
proof must rely on more than just empirical three-factor model from a Korean sample and
evidence (e.g., Bynner 1988; Galletta and a four-factor model was extracted from a Hong
Lederer 1989). Kong data set. The tangibles dimension was
retained in the analysis of both of the Asian
The results of several studies have demon- samples.
strated that the five dimensions claimed for the
SERVQUAL instrument are unstabie (see The unstable dimensionality of SERVQUAL
Tabie 2). SERVQUAL studies in the informa-
demonstrated in many domains, including
tion systems domain have also demonstrated
the unstable dimensionality of the SERVQUAL information services, is not just a statistical
instrument. The service quality of IS services curiosity. The scoring procedure for
was measured in three different industries, a SERVQUAL calls for averaging the P-E gap
financial institution, a consulting firm, and an scores within each dimension (Parasuraman et
information systems service business (Pitt et al. 1988). Thus a high expectation coupled
al. 1995). Factor analysis was conducted using with a iow perception for one item would be
principal components and maximum likelihood
canceled by a low expectation and high per-
methods with varimax rotation for a range of
ception for another item within the same
models. Analysis indicated differing factor
structures for each type of firm. Analysis of the dimension. This scoring method is oniy appro-
results for a financiai institution indicated a priate if all of the items in that dimension are
seven-factor modei with both the tangibles and interchangeable. This type of analysis would
empathy dimensions split into two. These be justified if SERVQUAL demonstrated a
breakdowns should not be surprising. Pitt et al. clear and consistent dimensional structure.
note that "up-to-date hardware and software" However, given the unstable number and pat-
are quite distinct from physical appearances in
tern of the factor structures, averaging groups
the IS domain. The empathy dimension was
created by the original SERVQUAL authors of items to calculate separate scores for each
from two distinctly different constructs, namely dimension cannot be justified. Therefore, for
understanding and access, which were com- scoring purposes, each item should be treated
bined due to the factor ioadings alone, without individually and not as part of some a priori
regard to underlying theory. Not only did IS- dimension.

MIS Quarteriy/June 1997 203


Research Note

sional collapsing utilized in the development of


SERVQUAL has resulted in a 22 paired-item
instrument that in fact does not measure all of
the theoretical dimensions of the service quali-
In summary, numerous problems with the origi- ty construct (i.e., content validity), then the use
nal SERVQUAL instrument are described in of linear sums of those items for purposes of
the literature (e.g., Babakus and Boiler 1992; measuring overall service quality is problemat-
Carman 1990; Cronin and Taylor 1992, 1994;
ic as well (Galletta and Lederer 1989).
Teas 1993, 1994). The evidence suggests that
difference scores, like the SERVQUAL percep- Many of the difficulties identified with the
tion-minus-expectation calculation, tend to SERVQUAL instrument also apply to the IS-
exhibit reduced reliability, poor discriminant modified versions of the SERVQUAL instru-
validity, spurious correlations, and restricted ment used by Pitt et al. (1995) and by
variance problems (Edwards 1995; Peter et al. Kettinger and Lee (1994). It appears that the
1993). The fact that the perception component IS-SERVQUAL instrument, utilizing difference
of the difference score exhibits better reliabili- scores, is neither a reliable nor a valid mea-
ty, convergent validity, and predictive validity surement for operationalizing the sen/ice quali-
than the perception-minus-expectation differ- ty construct for an information systems ser-
ence score itself calls into question the empiri- vices provider. The IS versions of the
cal and practical usefulness of both the expec- SERVQUAL instrument, much like the original
tations scores as well as the difference scores instrument (Parasuraman et al. 1988), suffer
(Babakus and Boiler 1992; Cronin and Taylor from unstable dimensionality and are likely to
1992; Parasuraman et al. 1994). Moreover, exhibit relatively poor predictive and conver-
inconsistent definitions and/or interpretations gent validity, as well as reduced reliability
of the "expectation" construct lead to a number when compared to non-difference scoring
of problems. The findings of Teas (1993) sug- methods. The existing literature provides
gest that a considerable portion of the variance impressive evidence that the use of percep-
in SERVQUAL scores is the result of measure- tion-minus-expectation difference scores is
ment error induced by respondents' varying problematic.
interpretations of the expectations construct. In
addition, since expectations, as well as per- This critique of the perceptions-minus-expecta-
ceptions, are subject to revision based on tions gap score should not be interpreted as a
experience, concerns regarding the temporal claim that expectations are not important or
reliability of SERVQUAL difference scores are that they should not be measured. Qn the con-
raised. Furthermore, the dimensionality of the trary, evidence indicates that both shouid and
SERVQUAL instrument is problematic. wiii expectations are precursors to perceptions
but that perceptions alone directly influence
It was reported that an analysis of IS- overall perceived sen/ice quality (Boulding et
SERVQUAL difference scores resulted in al. 1993). Our criticism is not with the concept
either three, five, or seven factors depending of expectations per se, but rather the opera-
on the industry (Pitt et al. 1995). A portion of tionalization of service quality as a simple sub-
the instability in the dimensionality of traction of an ambiguously defined expecta-
SERVQUAL can be traced to the development tions construct from the perceptions of the ser-
of the original instrument (i.e., Parasuraman et vice actually delivered.
al. 1988). Given these problems, users of the
SERVQUAL instrument should be cautioned to IS professionals have been known to raise
assess the dimensionality implicit in their spe- expectations to an unrealistically high level in
cific data set in order to determine whether the order to gain user commitment to new systems
hypothesized five-factor structure that has and technologies. This can make it much more
been proposed (Parasuraman et al. 1988, difficult to deliver systems and services that will
1991) is supported in their particular domain. be perceived as successful. According to the
Moreover, if the item elimination and dimen- model developed by Boulding et al. (1993), per-

204 MiS Quarterty/June 1997


SERVOUAL Concems

ceived service quality can be increased by expectations." Qne study indicates that such
either improving actual performance or by man- direct measures possess higher reliability and
aging expectations, specifically by reducing Improved convergent and predictive validity
should expectations and/or increasing will when compared to difference scores
expectations. These two different types of (Parasuraman et al. 1994a).
expectations are not differentiated by the tradi-
tional SERVQUAL gap scoring method. A better
approach to understanding the impact of expec-
tations on perceived service quality may be to
Conclusion
measure wiii and shou/af expectations separate-
Recognizing that we cannot manage what we
ly and then compare them to a service quality
cannot measure, the increasingly competitive
measure that utilizes either a direct response or
market for IS services has emphasized the
perceptions-only method of scoring.
need to develop valid and reliable measures of
the service quality of information systems ser-
vices providers, both internal and external to
Prescriptions for the use of the organization. An important contribution to
SERVQUAL this effort was made with the suggestion of a
IS-modified version of the SERVQUAL instru-
The numerous problems associated with the ment (Pitt et al. 1995). However, earlier stud-
use of difference scores suggest the need for ies raised several important questions con-
an alternative response format. Qne alterna- cerning the SERVQUAL instrument (e.g.,
tive is to use the perceptions-only method of Babakus and Boiler 1992; Carman 1990;
scoring. A review of the literature (Babakus Cronin and Taylor 1992, 1994; Peter et al.
and Boiler 1992; Boulding et al. 1993; Cronin 1993; Teas 1993, 1994). A review of the litera-
and Taylor 1992; Parasuraman et al. 1991, ture suggests that the use of difference scores
1994), indicates that perceptions-only scores with the IS-SERVQUAL instrument results in
are superior to the perception-minus-expecta- neither a valid nor reliable measure of per-
tion difference scores in terms of reliability, ceived IS service quality. Those choosing to
convergent validity, and predictive validity. In use any version of the IS-SERVQUAL instru-
addition, the use of perceptions-only scores ment are cautioned. Scoring problems aside,
reduces by 50% the number of items that must the consistently unstable dimensionality of the
be answered and measured (44 items to 22). SERVQUAL and IS-SERVQUAL instruments
Moreover, the findings of Boulding et al. (1993) intimates that further research is needed to
suggest that expectations are a precursor to determine the dimensions underlying the con-
perceptions and that perceptions alone directly struct of service quality. Given the importance
influence service quality. of the service quality concept in IS theory and
practice, the development of improved mea-
A second alternative, suggested by Carman sures of service quality for an information sys-
(1990) and Babakus and Boiler (1992), is to tems services provider deserves further theo-
revise the wording of the SERVQUAL items retical and empirical research.
into a format combining both expectations and
perceptions into a single question. Such an
approach would maintain the theoretical value
of expectations and perceptions in assessing References
service quality, as well as reduce the number
of questions to be answered by 50%. This Babakus, E., and Boiler, G. W. "An Empirical
direct response format holds promise for over- Assessment of the SERVQUAL Scale,"
coming the inherent problems with calculated Joumai of Business Research (24:3), 1992,
difference scores. Items with this format could pp. 253-268.
be presented with anchors such as "falls far Bagozzi, R., Davis, F., and Warshaw, P.
short of expectations" and "greatly exceeds "Development and Test of a Theory of

MIS Ouarterly/June 1997 205


Research Note

Technological Learning and Usage," Minus-Expectations Measurements of


Human Relations, (45:)7, 1992, pp. Service Ouality," Journai of Marketing
659-686. (58:1), 1994, pp. 125-131.
Baroudi, J. and Orlikowski, W. "A Short-Form Dabholkar, P. A., Thorpe, I. D., and Rentz, J.
Measure of User Information Satisfaction: A O. "A Measure of Service Ouality for Retail
Psychometric Evaluation and Notes on Stores: Scale Development and Validation,"
Use," Journai of Management information Journai of the Academy of Marketing
Systems (4:4), 1988, pp. 44-49. Sciences (24:1), 1996, pp. 3-16.
Boulding, W., Kaira, A., Staelin, R., and DeLone, W., and McLean, E. "Information
ZeithamI, V. A. "A Dynamic Process Model Systems Success: The Ouest for the
of Service Ouality: From Expectations to Dependent Variable," information Systems
Behavioral Intentions," Journai of Mari<eting Research (3:1), March 1992, pp. 60-95.
Research (30:1), 1993, pp. 7-27.
Doll, W. J., and Torkzadeh, G. "End-User
Brensinger, R. P., and Lambert, D. M. "Can
Computing Involvement: Discrepancy
the SERVQUAL Scale be Generalized to
Model," Management Science (35:10),
Business-to-Business Services?" in
1989, pp. 1151-1171.
Knowiedge Deveiopment in Mari<eting,
AMA's Summer Educators Conference Edwards. J. R. "Alternatives to Difference
Proceedings, Boston, MA, 1990, p. 289. Scores as Dependent Variables in the
Brown, T. J., Churchill, G. A., and Peter, J. P. Study of Congruence in Organizational
"Improving the Measurement of Service Research," Organizationai Behavior and
Ouality," Journai of Retaiiing (69:1), 1993, Human Decision Processes (64:3), 1995,
pp. 127-139. pp. 307-324.
Bynner, J. "Factor Analysis and the Construct Finn, D. W., and Lamb, C. W. "An Evaluation
Indicator Relationship," Human Reiations of the SERVOUAL Scales in a Retailing
(41:5), 1988, pp. 389-405. Setting," Advances in Consumer Research
Campbell, D. T., and Fiske, D. W. "Convergent (18), 1991, pp. 338-357.
and Discriminant Validation by the Galletta, D. F., and Lederer, A. L. "Some
Multitrait-Multimethod Matrix," Psy- Cautions on the Measurement of User
choiogicai Buiietin (56), 1959, pp. 81-105. Information Satisfaction," Decision
Carman, J. M. "Consumer Perceptions of Sciences (20:3), 1989, pp. 419-439.
Service Ouality: An Assessment of the Gronroos, C. "A Service Ouality Model and its
SERVOUAL Dimensions," Journai of Marketing Implications," European Joumai
Retaiiing (66:1), 1990, pp. 33-55. of Marketing (18:4), 1984, 36-55.
Cronbach, L. J. "Coefficient Alpha and the Johns, G. "Difference Score Measures of
Internal Structure of Tests," Psychometrii<a Organizational Behavior Variables: A
(16), 1951, pp. 297-334. Critique," Organizationai Behavior and
Cronbach, L. J., and Furby, L. "How We Human Performance (27), 1981, pp.
Should Measure 'Change'-or Should We?" 443-463.
Psychoiogicai Buiietin (74), July 1970, pp. Kappelman, L. "Measuring User Involvement:
74-73. A Diffusion of Innovation Approach," The
Cronbach, L. J., and Meehl, P. "Construct DATA BASE for Advances in information
Validity in Psychological Tests," Systems, (26:2,3), 1995, pp. 65-83.
Psychoiogicai Buiietin (21), 1955, pp. Kettinger, W. J., and Lee, C. C. "Perceived
281-302. Service Ouality and User Satisfaction with
Cronin, J. J., and Taylor, S. A. "Measuring the Information Services Function,"
Service Ouality: A Reexamination and Decision Sciences (25:5), 1994, pp.
Extension," Journai of Mari<eting (56:3), 737-766.
1992, pp. 55-68. Kettinger, W., Lee C, and Lee, S. "Global
Cronin, J. J., and Taylor, S. A. "SERVPERF Measurements of Information Service
versus SERVOUAL: Reconciling Ouality: A Cross-National Study," Decision
Performance-Based and Perceptions- Sciences (26:5), 1995, pp. 569-585.

206 MIS Ouarterly/June 1997


SERVQUAL Concems

Lord, F. M. "The Utilization of Unreliable Quality," Journai of Mari<eting (57:4), 1993,


Difference Scores," Journal of Educationai pp. 18-34.
Psychoiogy (49:3), 1958, pp. 150-152. Teas, R. K. "Expectations as a Comparison
Nunnaliy, J. Psychometric Theory, 2nd ed., Standard in Measuring Service Quality: An
McGraw-Hill, New York, 1978. Assessment of a Reassessment," Journai
Oliver, R. L. "A Cognitive Modei of the of Mari<eting (58:^), 1994, pp. 132-139.
Antecedents and Consequences of Van Dyke, T. P., and Popelka, M. E.
Satisfaction Decisions," Journai of "Development of a Quality Measure for an
Mari<eting Research (17:3), 1980, pp. Information Systems Provider," in The
460-469. Proceedings of the Decision Sciences
Parasuraman, A., ZeithamI, V. A., and Berry, institute (3), 1993, pp. 1910-1912.
L. L. "A Conceptual Model of Service Wall, T. D., and Payne, R. "Are Deficiency
Quality and its Implications for Future Scores Deficient?" Journai of Appiied
Research," Journal of Mari<eting (49:4), Psychoiogy (56:3), 1973, pp. 322-326.
1985, pp. 41-50. ZeithamI, V. A., Berry, L., and Parasuraman,
Parasuraman, A., ZeithamI, V. A., and Berry, A. "The Nature and Determinant of
L L. "SERVQUAL: A Multiple Item Scale for Customer Expectation of Service Quality,"
Measuring Consumer Perceptions of Journai of the Academy of Mariteting
Service Quality," Journai of Retaiiing (64:1), Science (2V.\), 1993, pp. 1-12.
1988, pp. 12-40.
Parasuraman, A., ZeithamI, V. A., and Berry,
L. L. "Refinement and Reassessment of the About the Authors
SERVQUAL Scale," Journai of Retaiiing
(67:4), 1991, pp. 420-450. Thomas P. Van Dyke is an assistant profes-
Parasuraman, A., ZeithamI, V. A., and Berry, sor of information systems and technologies at
L. L. "Alternative Scales for Measuring Weber State University and recently complet-
Service Quality: A Comparative ed his a doctoral dissertation in business com-
Assessment Based on Psychometric and puter information systems at the University of
Diagnostic Criteria," Journai of Retaiiing North Texas. His current research interests
(70:3), 1994a, pp. 201-229. include the effects of alternative presentation
Parasuraman, A., ZeithamI, V. A., and Berry, formats on biases and heuristics in human
L. L. "Reassessment of Expectations as a decision making and MIS evaluation and
Comparison in Measuring Service Quality: assessment. He has published articles in The
Implications for Further Research," Journai Journai of Computer information Systems and
ofMari<eting(5a:^), 1994b, pp. 111-124. Proceedings of the Decision Sciences
Peter, J. P., Churchill, G. A., and Brown, T. J. institute.
"Caution in the Use of Difference Scores in
Consumer Research," Journai of Consumer Leon A. Kappelman is an associate professor
Research (19:1), 1993, pp. 655-662. of business computer information systems in
Pitt, L. F., Watson, R. T., and Kavan, C. B. the College of Business Administration at the
"Service Quality: A Measure of Information University of North Texas and associate direc-
Systems Effectiveness," MiS Ouarteriy tor of the Center for Quality and Productivity.
(19:2), June 1995, pp. 173-187. After a successful career in industry, he
Prakash, V., and Lounsbury, J. W. "A received his Ph.D. (1990) in management
Reliability Problem in the Measurement of information systems from Georgia State
Disconfirmation of Expectations," in University. He has pubiished over two dozen
Advances in Consumer Research, Vol. 10, journal articles. His work has appeared in
Tybout, A.M. and Bagozzi, R. P. (eds.) Communications of the ACM, Journai
Association for Consumer Research, Ann of Management information Systems,
Arbor, Ml, 1983, pp. 244-249. DATA BASE for Advances in information
Teas, R. K. "Expectations, Performance Systems, Journai of Systems Management,
Evaluation and Consumer's Perception of Journai of Computer information Systems,

MIS Quarteriy/June 1997 207


Research Note

InformationWeek, National Productivity and Productivity and an associate professor of


Review, Project Management Joumal, Joumal management science in the Business
of Information Technology Management,
Computer Information System Department. He
Industrial Management, as well as other jour-
has published articles in Technometrics,
nals and conference proceedings. He authored
Information Systems for Managers, McGraw- Operations Research, Economic Quality
Hill (1993). His current research interests Control, and Ouaiity Progress, as well as other
include the management of information assets, journals and conference proceedings. Dr.
information systems (IS) development and Prybutok is a senior member of the American
implementation, IS project management, and Society for Quaiity Control (ASQC), an ASQC
MIS evaluation and assessment. He is co- certified quality engineer, a certified quality
chair of the Society for Information
auditor, and a 1993 Texas quaiity award
Management's (SIM) Year 2000 Working
Group. examiner. His current research interests
include project management, assessment of
Victor R. Prybutok is the director of the quality programs, neural networks, and MIS
University of North Texas Center for Quality evaluation and assessment.

208 MIS Quarteriy/June 1997


View publication stats

You might also like