Professional Documents
Culture Documents
IJSIM
15,3 Online service quality dimensions
and their relationships with
satisfaction
302
A content analysis of customer reviews of
securities brokerage services
Zhilin Yang
Department of Marketing, City University of Hong Kong, Kowloon,
Hong Kong, and
Xiang Fang
Department of Marketing, Oklahoma State University, Tulsa, Oklahoma, USA
Keywords Electronic commence, Internet, Servicing, Customer satisfaction, Securities markets
Abstract This exploratory research intends to extend our understanding of service quality and
customer satisfaction within the setting of online securities brokerage services. Based upon
conceptual frameworks from the areas of services marketing and information systems
management, the authors uncovered 52 items across 16 major service quality dimensions by
content analysis of 740 customer reviews. The results indicate that primary service quality
dimensions leading to online customer satisfaction, with the exception of ease of use, are closely
related to traditional services while key factors leading to dissatisfaction are tied to information
systems quality. In addition, major drivers of satisfaction and dissatisfaction are identified at the
sub-dimensional level. Theoretical contribution and managerial implications of the findings are
further discussed.
At least once a week it’s impossible to logon, get real-time quotes, or have my portfolio quotes
displayed. In some instances, I couldn’t execute trades or check under key shortfalls for this
company . . . (Quote from one online investor’s comment at www.gomez.com).
Web-based services (hereinafter online services or e-services) offer customers a
panoply of benefits such as enhanced control, ease of use, and reduced transaction
charges (Scullion and Nicholas, 2001; Unsal and Movassaghi, 2001; Zeithaml, 2002).
Consequently, online services have grown rapidly and have emerged as the leading
edge of the service industry. For instance, in the USA, the number of online investors
totaled 17.4 million as of the year 2001; and this user level is predicted to reach 32.4
million by 2004 (eMarketer, 2002). To accommodate this growth, not only have dozens
of Internet-only service providers materialized, but also significant numbers of
conventional suppliers have utilized the Internet to furnish online services. The
consequence of this phenomenon has been the appearance of ferocious competition
International Journal of Service among online suppliers.
Industry Management
Vol. 15 No. 3, 2004
pp. 302-326 The first author gratefully acknowledges a research grant from the City University of Hong
q Emerald Group Publishing Limited
0956-4233
Kong (SMA Project No. 9030907). The authors would like to express our thanks to Nancy Wong,
DOI 10.1108/09564230410540953 Wing Yeung, and Yue Shen for their research assistance.
One effective means of achieving competitive advantage and differentiating Online service
strategies involves superior service quality. Superior service quality measurably quality
increases a firm’s overall profitability, its price premium, and its perceived and actual
market share (Zeithaml et al., 1996). Most online service providers (particularly
securities brokers), however, have encountered substantial problems and challenges in
furnishing online service quality. As confirmed by Goldberg (2000, p. 42):
Many investors had difficulty making trades because Web sites crashed or were glacially
303
slow. Clients with problems or questions were kept on perpetual hold. E-mails weren’t
answered.
The primary reasons for these difficulties reside in service providers’ lack of experience
in streamlining operations of this fairly recent channel and their limited understanding
of online customers’ usage behaviors. Moreover, as online firms often possess limited
resources, priorities must therefore be determined among alternative service attributes
when making investment decisions related to achievement of superior services
(Brandt, 1997). As such, it is desirable for online service providers to uncover what
attributes consumers utilize in their assessment of overall service quality and
satisfaction and which attributes are more important.
From the theoretical perspective, research in e-services is still in its early stages
although e-commerce has experienced explosive growth during the past decade. Online
firms operate in an environment consisting of Web-based technologies (e.g. browsers,
search engines, encryption, and databases), network systems, and digital information.
This unique feature of online services has motivated us to re-examine whether
traditional service quality dimensions and their content are truly applicable to
Internet-based services and to explore new quality dimensions (Cox and Dale, 2001).
Few scholarly studies, to date, have been undertaken to identify quality dimensions
and detailed aspects of online services and their relationships with customer
satisfaction (Zeithaml, 2002, Zeithaml and Bitner, 2003). Moreover, in current studies of
online service quality, “scholars too often move directly from the generation of items to
factor analysis, thus giving limited considerations to how consumers categorize and
think about their experiences” (Wolfinbarger and Gilly, 2002, p. 4).
Thus, our exploratory research intends to extend the body of existing knowledge
relating to service quality and customer satisfaction in multiple ways. First, we
establish a conceptual framework by integrating two major aspects of online services
– customer service quality and information systems quality. Based upon the
conceptual framework, we employ a content analysis of customer reviews on online
securities brokerage services and then identify 16 major quality dimensions of online
services. We further uncover 52 sub-dimensions or items within all major dimensions.
Third, we examine primary service quality dimensions leading to satisfaction and
dissatisfaction respectively. The results suggest that primary drivers of satisfaction
are closely related to traditional service quality dimensions with the exception of “ease
of use”, and key drivers of dissatisfaction are tied to information systems quality. In
addition, we extend Johnston’s (1995) classification of satisfiers and dissatisfiers to the
sub-dimensional level and identify the major drivers of satisfaction and dissatisfaction.
Fourth, we utilize a new approach, content analysis of 740 online customer reviews,
which enables us to identify relative comprehensive service quality dimensions and
sub-dimensions and to investigate their relationships with satisfaction. Finally, our
IJSIM research has been conducted in the context of online securities brokerage to which
15,3 virtually no empirical research, to the best of our knowledge, has been directed.
Conceptual development
Our conceptual framework has been developed on the basis of two important
characteristics of online services. First, online services invariably share some common
304 territory in relation to traditional interpersonal services. Second, Web-based
technologies have been used to automate product distribution and customer
services, including transaction and payment systems, call centers, customer
relationship management systems, as well as the underlying analytics, reporting,
and operations of these systems. In this sense, Web sites are functioning as
well-defined information systems. These two aspects are indeed overarching; they
interact with each other. Customer services may be delivered partially or even fully
through the Internet while maintaining information systems quality often requires the
aid of human beings. In line with this reasoning, we briefly review major literature on
traditional customer services quality, information systems quality applicable to online
services, and a combination of the two (i.e. online service quality).
Wolfinbarger and Gilly (2002), through focus group interviews and an online survey,
have uncovered four online retailing experience factors: Web site design, reliability,
privacy/security, and customer service. Yang and Jun (2002) have uncovered six
prominent factors – reliability, access, ease of use, personalization, security, and
credibility – employed by Internet purchasers to evaluate e-tailers’ service quality and
seven principal service attributes – security, responsiveness, ease of use, reliability,
availability, personalization, and access – that could prevent visitors from purchasing.
These retailing services oriented studies have certain limitations in terms of the
generalization of their findings to other settings.
Some studies develop online service quality framework based on literature review.
Cox and Dale (2001) have noted that traditional service quality dimensions, such as
competence, courtesy, cleanliness, comfort, and friendliness, are not relevant to online
retailing; whereas other factors, such as accessibility, communication, credibility, and
appearance, are critical to its success. Madu and Madu (2002) proposed the following
15 dimensions of online service quality based on literature review: performance,
features, structure, aesthetics, reliability, storage capacity, serviceability, security and
system integrity, trust, responsiveness, product/service differentiation and
customization, Web store policies, reputation, assurance, and empathy. Zeithaml
et al. (2002) conducted a critical review of extant e-service quality literature and
identified seven dimensions for e-service quality – efficiency, reliability, fulfillment,
privacy, responsiveness, compensation, and contact. These seven dimensions were
IJSIM further classified into two categories: a core service scale for measuring the customers’
15,3 perceptions of e-tailers’ service quality when customers experience no problems with
the site (efficiency, reliability, fulfillment, and privacy) and a recovery service scale for
measuring e-tailers’ service recovery when problems occur (responsiveness,
compensation, and contact).
Web site design quality. From the information systems perspective, Yoo and Donthu
308 (2001) have developed an instrument for measuring online retailers’ Web sites,
SITEQUAL, with four dimensions: ease of use, aesthetic design, processing speed, and
security. Another highly cited instrument, WEBQUAL (Web quality) developed by
Loiacono et al. (2002) consists of 12 dimensions: information fit to task, interaction,
trust, response time, design, intuitiveness, visual appeal, innovativeness, flow,
integrated communication, business process, and substitutability. Liu and Arnett
(2000), through Web and e-mail surveys of 689 Webmasters of Fortune 1,000
companies’ Web site, have discovered four major factors leading to the success of a
Web site, i.e. information and service quality, system use, playfulness, and system
design quality. These scales, however, have one major limitation as they are “geared
toward helping Web site designers to better design Web sites to affect the interaction
perceptions of the users. Therefore, the scale is more pertinent to interface design than
service quality measurement” (Zeithaml et al., 2002, p. 366).
Other online service quality. Other studies have identified key dimensions of online
service quality in the context of narrowly defined online businesses, such as portal
services and travel agencies. In each of these online services, some different dimensions
emerged. For instance, Van Riel et al. (2001) have derived three key portal service
quality attributes affecting consumers’ overall satisfaction with the portal: core service,
support service, and user interface. Kaynama and Black (2000) have identified the
following dimensions for online travel agencies: content and purpose, accessibility,
navigation, design and presentation, responsiveness, background, and personalization
and customization.
The foregoing literature review suggests that most of previous empirical studies
have focused either on service quality of retailing-oriented Web sites (e.g. Yang and
Jun, 2002; Zeithaml et al., 2002), or on the technical quality of Web site itself (e.g. Liu
and Arnett, 2000; Loiacono et al., 2002). Little research has integrated models from
traditional service quality and information systems quality; and consequently, the
unique quality aspects of online services (particularly those from service-intensive
industries) could not be fully reflected. In addition, the very few exploratory studies on
online service quality have utilized focus groups and personal interviews. These
research methods usually have very limited numbers of participants, which prevents
them from uncovering more comprehensive dimensions and their sub-dimensions (e.g.
Van Riel et al., 2001). These limitations have motivated us to conduct an exploratory
study to address the following two fundamental questions:
RQ1. Which dimensions and sub-dimensions of online services assume vital roles
in influencing customers’ evaluation of service quality?
RQ2. Compared to traditional services, which service quality dimensions and/or
sub-dimensions are considered specific to online services?
Service quality dimensions and customer satisfaction Online service
Researchers have paid much attention to the close relationship between service quality quality
and customer satisfaction (e.g. Bitner et al., 1990; Parasuraman et al., 1988). Oliver
(1993) suggests that service quality is a more specific judgment, which can lead to a
broad evaluation, customer satisfaction. The question is: how exactly will particular
service quality dimensions influence customer satisfaction formation? Johnston (1995)
has found that the causes of dissatisfaction and satisfaction are not necessarily the 309
same. Some service quality attributes may not be critical for customer satisfaction but
can significantly lead to dissatisfaction when they are performed poorly. Drawing
upon Herzberg et al.’s (1995) research on work motivation, Johnston (1997) has further
classified all dimensions into enhancing (satisfiers), hygiene (dissatisfiers), and dual
factors. Enhancing factors are those which will lead to customer satisfaction if they are
delivered properly, but will not necessarily cause dissatisfaction if absent. In contrast,
hygiene factors are those which will lead to customer dissatisfaction if they fail to
deliver, but will not result in satisfaction if present. Dual factors are those that will
have an impact on both satisfaction and dissatisfaction. Johnston (1995) identified
attentiveness, responsiveness, care, and friendliness as the main source of satisfaction
in banking services, and integrity, reliability, responsiveness, availability, and
functionality as the main source of dissatisfaction.
In addition, Mittal et al. (1998) have discovered that the positive performance of a
service quality attribute has less impact on overall consumer satisfaction than negative
performance of that same attribute. This “asymmetric effect” suggests that it is
necessary to carefully investigate service quality attributes having both negative and
positive contributions, and to grant more attention to those salient
negatively-performed attributes to increase customer satisfaction. Is this asymmetric
effect applicable to online services? Are there any differences between online services
and traditional services? Thus, the third key question we intend to address is:
RQ3. Among these identified service quality dimensions, which are principally
leading to online customer satisfaction (“satisfiers”) and which are leading to
online dissatisfaction (“dissatisfiers”)?
Research method
Consumer feedback, e.g. complaints and compliments have long been employed as
sources of feedback on company performance. Essentially, customer feedback is an
exception-reporting mechanism for identifying weaknesses to be corrected or strengths
to be reinforced. The fact that customers contribute time and effort for voicing their
online service experiences suggests that the attributes are salient in the post-use
evaluation process (Cadotte and Turgeon, 1988). Although consumer comments are not
likely to reflect completely their entire experiences with suppliers, they do highlight
those dimensions and detailed attributes of service quality of greatest concern (Yang
and Peterson, 2002; Yang et al., 2001). Therefore, an analysis of consumer compliments
and complaints are particularly useful in exploring the rich aspects of customer
perceived service quality (Yang and Peterson, 2003). Unfortunately, a review of the
consumer behavior literature reveals that consumer’s anecdotal feedback of online
services has not been carefully examined, and deserves more academic attention.
Our study, therefore, adopted the newly emerged “netnography” method by
utilizing a content analysis of consumer reviews of online brokerage experiences to
identify key dimensions and their respective service features. The “netnography”
approach is a qualitative research technique which employs an ethnographic research
method to study online customers (Kozinets, 2002). It “uses the information publicly
available in online forums to identify and understand the needs and decision influences
of relevant online consumer groups” (Kozinets, 2002, p. 62). Online content analysis is
part of “netnography” in the sense that it is based on content created by online
customers and intends to understand their needs and wants. Compared to content
generated by traditional methods (e.g. focus groups and personal interviews),
“netnography” is less-time consuming, less obtrusive, less costly; more timely, natural,
and elaborate. “The limitations of ‘netnography’ draw from its more narrow focus on
online communities, the need for researcher interpretive skill, and the lack of informant
identifiers present in the online context that leads to difficulty generalizing community
sample” (Kozinets, 2002, p. 62).
Sample Online service
We employed three steps to collect qualified customer reviews of brokerage services. quality
The first step was to locate appropriate Web sites that provide customers with a hub to
evaluate suppliers. By using multiple search sites (i.e. Google, Altavista, MSN search,
LookSmart, and Hotbot), we intensively reviewed the most prominent US-based Web
sites that offer online consumer reviews. Three principal criteria were set to select
appropriate sites: the site should be established by a party independent of product 311
suppliers; the site should have a review section on online securities brokerage services;
and reviews should be written by customers who have used or are using the service
and have been authorized to post their messages (Yang and Peterson, 2003). We then
found that seven Web sites were relevant for this study – bizrate.com, complaints.com,
computingreview.com, consumerreports.org, epinions.com, gomez.com, and rateitall.
com
The second step was to select qualified customer review sites. We established three
selection criteria to permit collection of the most representative samples: the number of
reviews written by online investors should be large; customers should be encouraged
to post both dissatisfied and satisfied reviews; and the reviews can be assessed
publicly or freely. Among the above-mentioned seven Web sites, five did not satisfy
our criteria. Consumerreports.org did not allow visitors to view customer reviews
freely. The number of customer reviews contributed by diversified participants in
Bizrate.com, rateitall.com, and computingreview.com was too small. Finally,
Complaints.com had limited reviews of online brokers and exclusively consisted of
complaint messages. The remaining two Web sites, epinion.com, a leading online
product review site, and gomez.com, a leading customer review site dedicated to online
financial services, fully met the requirements.
The last step was to choose sample Internet securities brokers. Five of the most
influential online brokers were selected for study – Fidelity Investment, Charles
Schwab, Ameritrade, TD Waterhouse, and E-Trade. These online brokers were ranked
as the top five within 21 leading online brokers according to CNN (2001). While
Ameritrade and E-Trade are Internet-only firms, the remaining three firms are hybrid
(or click-and-brick) brokers.
Data collection
We accessed the two online customer review Web sites, Gomez and Epinions, from
January 12 to 15, 2002, to secure a sufficient volume of anecdotes. All messages written
from January 1, 2000, to January 11, 2002 were collected. We made every effort in
deleting disqualified reviews, e.g. spamming messages, personal attacks, duplications,
messages with empty text body and a meaningless title, promotional messages written
by company insiders, and any messages irrelevant to online brokerage services. We
further identified multiple messages written by the same person for the same broker
and chose the message that best represented his/her viewpoints. As a result, a total of
740 effective consumer anecdotes were selected. The average length of each anecdote
was 124 words.
Coding process
Our coding procedure followed the guideline provided by Yang and Peterson (2002).
All anecdotes were numbered, formatted, and imported to Ethnograph 5.0, a software
IJSIM package designed for coding qualitative data (Wazienski, 2000). Based on the first 100
15,3 messages, two researchers collaborated in developing an initial 59 coding words, which
made up the primary themes or facets of the overall quality of online services.
Whenever possible, we grouped these coding words into relevant factors defined by
previous studies (e.g. Doll and Torkzadeh, 1988; Johnston, 1995; Parasuraman et al.,
1988). Since some coding words could not be assigned to the factors uncovered by
312 major extant studies, new dimensions were subsequently developed. The two
researchers then independently coded the remaining anecdotes. Subsequent discussion
identified and resolved for all disagreements. Based on reviewers’ attitude, each service
quality dimension was classified into two categories: satisfied attributes (positive
performance) and dissatisfied attributes (negative performance).
The inter-judge reliability between the coders was calculated using the percentage
agreement statistic. In turn, the inter-judge reliability was 88.5 percent for the
satisfying anecdotes and 84.3 percent for the dissatisfying anecdotes. These figures are
relatively high, given that the resulting classification scheme contained 16 categories
according to standards normally employed for statistically assessing qualitative data
(Perreault and Leigh, 1989).
Results
The results of content analysis are reported in Tables I and II. We present findings in
line with our three research questions. Whenever appropriate, illustrative anecdotes
are quoted to explain the domain of service quality dimensions and sub-dimensions.
1. Responsiveness Prompt response to phone calls 0.0 118 11.5 118 9.7
Prompt services (others) 22 11.8 64 6.2 86 7.1
Prompt response to e-mails 0.0 54 5.2 54 4.4
314 Prompt order execution 39 20.9 66 6.4 105 8.6
Prompt order confirmation 6 3.2 17 1.7 23 1.9
Sub-total 67 35.8 319 31.0 386 31.7
2. Service reliabilty Accurate order fulfillment 12 6.4 62 6.0 74 6.1
Accurate record 2 1.1 18 1.7 20 1.6
Accurate quote 6 0.6 6 0.5
Accuracy in billing 3 1.6 3 0.2
Accurate calculation of commissions 3 0.3 3 0.2
Keep service promise 27 2.6 27 2.2
Keep promotion promise 7 0.7 7 0.6
Sub-total 17 9.1 123 12.0 140 11.5
3. Ease of use Easy navigation 4 2.1 11 1.1 15 1.2
Easy to use 8 4.3 6 0.6 14 1.2
Effective navigation 5 0.5 5 0.4
Functionality/interface 8 0.8 8 0.7
Web site download speed 7 3.7 86 8.4 93 7.6
Sub-total 19 10.2 116 11.3 135 11.1
4. Competence Ability to solve problems 17 9.1 39 3.8 56 4.6
Knowledge to answer questions 5 2.7 27 2.6 32 2.6
Research capacity 6 3.2 25 2.4 31 2.5
Quickly solve problem 2 1.1 15 1.5 17 1.4
Sub-total 30 16.0 106 10.3 136 11.2
5. Access Web site accessibility 65 6.3 65 5.3
Phone access 29 2.8 29 2.4
Local office access 6 3.2 6 0.5
E-mail access 5 0.5 5 0.4
Sub-total 6 3.2 99 9.6 105 8.6
6. System reliability System failure and shutdown 49 4.8 49 4.0
System error 31 3.0 31 2.5
Sub-total 0 0.0 80 7.8 80 6.6
7. Timeliness Account update 2 1.1 41 4.0 43 3.5
Quote update 4 2.1 27 2.6 31 2.5
Sub-total 6 3.2 68 6.6 74 6.1
8. Security Log off due to time-out 22 2.1 22 1.8
Lock out from service 7 0.7 7 0.6
Personal information privacy 4 0.4 4 0.3
Customer protection 1 0.5 1 0.1
Fund or trade hold-up 6 0.6 6 0.4
Table III. Sub-total 1 0.5 39 3.8 40 3.3
Frequencies of service Total (eight dimensions)
quality dimensions, by 146 78.1 950 92.3 1096 90.1
Grant total (16 dimensions) 187 100.0 1,029 100.0 1216 100.0
satisfaction and
dissatisfaction Note: a Percentage is calculated based upon a total of 1,216 comments (16 dimensions)
groups. They are: responsiveness (a mention quotient of 31.7 percent), service Online service
reliability (11.5 percent), ease of use (11.1 percent), competence (11.2 percent), access quality
(8.6 percent), system reliability (6.6 percent), timeliness (6.1 percent), and
security/privacy (3.3 percent). These eight factors account for 90.1 percent of
remarks while the remaining eight dimensions consist only of 9.9 percent. In the same
vein, the top eight items or keywords customers mentioned are prompt response to
phone calls (9.7 percent), prompt order execution (8.6 percent), Web site download 315
speed (7.6 percent), prompt customer services (7.1 percent), accurate order fulfillment
(6.1 percent), Web site accessibility (5.3 percent), and representatives’ capability to
solve problems (4.6 percent).
The fact that the less-frequently mentioned eight factors appear in only about 10
percent of remarks obtained deserves further explanation. If the reason for such a low
number of citations is that the overall performance of these factors is in line with
customers’ expectations and/or their performance deviation from the industry
standard is insignificant, the less-frequently mentioned attributes may still be critical
for customer satisfaction. In this case maintaining high performance of these attributes
is a necessity for competition; but doing so may not gain competitive advantages. If, on
the other hand, the reason these factors are cited so infrequently is that they play a less
important role in customers’ evaluation of overall service performance, management
may strategically maintain only acceptable quality levels of these attributes and devote
more effort and resources on the other eight attributes in order to achieve competitive
advantages.
Figure 1.
The sources of satisfaction
and dissatisfaction for all
16 dimensions
courtesy, service portfolio, and continuous improvement can please online investors; Online service
their poor performance, nevertheless, is not the major reason of complaints. Similarly, quality
underperformance of access, system reliability, or timeliness tends to lead to
dissatisfaction: however, their adequate performance may not delight customers since
customers regarded these factors as necessities. The results are consistent with
Johnston’s (1997) finding that some determinants of service quality tend to lead to
satisfaction while others lead to dissatisfaction. 319
It is noteworthy that Johnston (1995) adopted the concept of “satisfiers” and
“dissatisfiers” only at the level of service quality dimensions. Our content analysis
enables us to take a further step in applying this notion to the sub-dimensional level of
quality dimensions. Figure 2 presents the satisfied and dissatisfied mention quotients
for some important items (. 1 percent) within four key service quality factors. For
“service reliability”, accurate order fulfillment is the most frequently mentioned item
leading to both satisfaction and dissatisfaction. For “competence”, ability to solve
problems is the main cause of both satisfaction and dissatisfaction. For the other two
factors, the major sources of satisfaction and dissatisfaction are different. For instance,
among all items of “responsiveness”, the mostly mentioned one for complaints is “long
waiting for phone responses”. However, prompt response to customer calls seems not
to be the major contributor of customer compliments. The main source of satisfaction is
prompt order execution. For “ease of use”, the main drivers of satisfaction is “easy to
use” and the main source dissatisfaction is “Web site download speed”. This finding
suggests that examining what leads to customer satisfaction and dissatisfaction
Figure 2.
The sources of satisfaction
and dissatisfaction for
some sub-dimensions
(percentage . 1 percent)
IJSIM should be conducted at both dimensional- and sub-dimensional (item)-level in order to
15,3 generate more insightful results.
Discussion
Listening to customer voices
Listening to customer voices is the initial step in planning service quality improvement
320 endeavors. In turn, identifying customer perceived service quality dimensions and
their roles in customer satisfaction and dissatisfaction provides a frame of reference for
online service providers to assess their service performance. Our online content
analysis uncovers 16 dimensions and 52 related items concerned with the role of online
service firms in facilitating clients’: choice of the right products by providing adequate
and timely information; speedy and accurate completion of a transaction; and
satisfactory resolution of any problems incurred. Once this preliminary activity has
been satisfactorily completed, full-fledged improvement efforts can be initiated.
Ideally, management should focus attention on the full spectrum of 16 dimensions.
Given the constraints of organizational resources, priority, however, has to be granted
to the following key dimensions: responsiveness, service and system reliability, ease of
use, timeliness, access, competence, and security, reflecting the mention frequency of
attributes with negative performance. These critical dimensions embrace factors from
both traditional service quality and information systems quality, which is consistent
with the unique aspects of Internet-based services. In addition, incorporating the items
specific to the Internet has enriched the content of uncovered factors.
Conclusion
In this study, we develop a broad conceptual framework integrating well-established
323
models from the areas of service marketing and information systems. We then employ
the newly emerged “netnography” method by content analysis of 740 anecdotes
written by online customers. The method enables us to uncover 16 dimensions and 52
related items, and their role in customer satisfaction and dissatisfaction. The findings
provide both theoretical contributions to the knowledge body of online service quality
and practical implications for online service providers.
Theoretically, this study extends the knowledge body of service quality and
customer satisfaction by enriching the content of traditional service quality and
information systems quality dimensions applicable for Internet-enabled services
and identifying multiple new factors. Compared to previous studies, the
dimensions and their related items developed in this research comprise a
relatively comprehensive pool of measures for assessing online services and can
serve as building blocks for further studies in relationship marketing. We further
reveal that major drivers of online service satisfaction are still strongly tied to
traditional service quality dimensions (e.g. responsiveness, service reliability, and
competence) although they contain some Internet-related contents. In contrast, the
primary determinants of dissatisfaction tend to relate to information and systems
quality dimensions. Among the factors unique to the Internet setting, ease of use
is the only major driver of both customer satisfaction and dissatisfaction.
In practice, subtle differentiation of service quality levels has become a key driving
force in enhancing competitive advantages. If online providers understand what
dimensions customers utilize to judge quality and form their satisfaction, they will then
be in a favorable position to monitor and improve company performance. In this sense,
the service quality dimensions and sub-dimensions identified in this study can
facilitate firms in detection of the weaknesses and strengths of their online services.
Management can thus devote valuable corporate resources to enhance performance of
salient service quality attributes such as service reliability, responsiveness, ease of use,
and competence identified by this study. Finally, while superior information and
systems quality may not delight customers, management still needs to maintain
quality levels that adhere to industry standards and should remedy any failure of these
attributes to avoid detrimental effects on customer satisfaction.
References
Adams, D.A., Nelson, R.R. and Todd, P.A. (1992), “Perceived usefulness, ease of use, and usage of
information technology: a replication”, MIS Quarterly, Vol. 16 No. 2, pp. 227-47.
Berry, L.L., Zeithaml, V.A. and Parasuraman, A. (1985), “Quality counts in services, too”,
Business Horizons, Vol. 28 No. 3, pp. 44-52.
Bienstock, C.C. (1997), “Measuring physical distribution service quality”, Journal of the Academy
of Marketing Science, Vol. 25 No. 1, pp. 31-44.
IJSIM Bitner, M.J., Booms, B.H. and Tetreault, M.S. (1990), “The service encounter: diagnosing
favorable and unfavorable”, Journal of Marketing, Vol. 54, January, pp. 71-84.
15,3
Brandt, R. (1997), “Satisfaction studies must measure what a customer wants and expects”,
Marketing News, Vol. 31, p. 17.
Cadotte, E.R. and Turgeon, N. (1988), “Dissatisfiers and satisfiers: suggestions for consumer
complains and compliments”, Journal of Consumer Satisfaction, Dissatisfaction and
324 Complaining Behavior, Vol. 1, pp. 74-9.
CNN (2001), “CNN Money Survey”, available at: http://money.cnn.com/best/brokers/index.html
(accessed 10 January 2002).
Cox, J. and Dale, B.G. (2001), “Service quality and e-commerce: an exploratory analysis”,
Managing Service Quality, Vol. 11 No. 2, pp. 121-31.
Davis, F.D. (1989), “Perceived usefulness, perceived ease of use, and user acceptance of
information technology”, MIS Quarterly, Vol. 13 No. 3, pp. 319-40.
Davis, F.D., Bagozzi, R.P. and Warshaw, P.R. (1989), “User acceptance of computer technology: a
comparison of two theoretical models”, Management Science, Vol. 35 No. 8, pp. 982-1003.
DeLone, W.H. and McLean, E.R. (1992), “Information system success: the quest for the dependent
variable”, Information Systems Research, Vol. 3 No. 1, pp. 60-95.
Doll, W.J. and Torkzadeh, G. (1988), “The measurement of end-user computing satisfaction”, MIS
Quarterly, Vol. 12 No. 2, pp. 259-74.
Doll, W.J., Xia, W. and Torkzadeh, G. (1994), “A confirmatory factor analysis of the end-user
computing satisfaction instrument”, MIS Quarterly, Vol. 18 No. 4, pp. 43-461.
eMarketer (2002), “Online investing: brokers, investors, statistics, and market trends”, available
at: www.emarketer.com/products/report.php?2000106 (accessed 21 April 2003).
Goldberg, S.T. (2000), “Online trading’s agony and ecstasy”, Investing Kiplinger’s, October,
pp. 41-7.
Hedvall, M.-B. and Paltschik, M. (1989), “An investigation in and the generation of service quality
concepts”, in Avlonitis, G.J., Papavasiliou, K. and Kouremenos, A.G. (Eds), Marketing
Thought and Practices in the 1990s, European Marketing Academy, Athens, pp. 473-83.
Hendrickson, A.R. and Collins, M.R. (1996), “An assessment of structure and causation of is
usage”, The Database for Advances in Information Systems, Vol. 27 No. 2, pp. 61-7.
Herzberg, F., Mausner, B. and Snyderman, B. (1995), The Motivation to Work, Wiley, New York,
NY.
Igbaria, M., Zinatelli, N., Cragg, P. and Cavaye, A.L.M. (1997), “Personal computing acceptance
factors in small firms: a structural equation model”, MIS Quarterly, Vol. 21 No. 3,
pp. 279-301.
Johnston, R. (1995), “The determinants of service quality: satisfiers and dissatisfiers”,
International Journal of Service Industry Management, Vol. 8 No. 5, pp. 53-71.
Johnston, R. (1997), “Identifying the critical determinants of service quality in retail banking:
importance and effect”, International Journal of Bank Marketing, Vol. 15 No. 4, pp. 111-6.
Kaynama, S.A. and Black, C.I. (2000), “A proposal to assess the service quality of online travel
agencies: an exploratory study”, Journal of Professional Services Marketing, Vol. 21 No. 1,
pp. 63-88.
Kettinger, W.K. and Lee, C.C. (1997), “Pragmatic perspectives on the measurement of information
systems service quality”, MIS Quarterly, Vol. 21, June, pp. 223-40.
Kozinets, R.V. (2002), “The field behind the screen: using Netnography for marketing research in
online communities”, Journal of Market Research, Vol. 39, February, pp. 61-72.
Lin, C.S. and Wu, S. (2002), “Exploring the impact of online service quality on portal site usage”, Online service
Proceedings of the 35th Hawaii International Conference on System Science.
quality
Lin, X. and Wei, B. (1999), “Service quality dimensions of securities brokerage firms: what
customers consider as important”, Journal of Professional Services Marketing, Vol. 20 No. 1,
pp. 135-46.
Liu, C. and Arnett, K.P. (2000), “Exploring the factors associated with Web site success in the
context of electronic commerce”, Information & Management, Vol. 38, October, pp. 23-34. 325
Loiacono, E.T., Watson, R.T. and Goodhue, D.L. (2002), “WEBQUAL: a measure of Web site
quality”, 2002 Marketing Educators’ Conference: Marketing Theory and Applications,
Vol. 13, pp. 432-7.
Madu, C.N. and Madu, A.A. (2002), “Dimensions of e-quality”, International Journal of Quality &
Reliability Management, Vol. 19 No. 3, pp. 246-58.
Mehta, S.C., Lalwani, A.K. and Han, S.L. (2000), “Service quality in retailing: relative efficiency of
alternative measurement scale for different product-service environments”, International
Journal of Retail & Distribution Management, Vol. 28 No. 2, pp. 62-72.
Mittal, V., Ross, W.T. and Baldasare, P.M. (1998), “The asymmetric impact of negative and
positive attribute-level performance on overall satisfaction and repurchase intentions”,
Journal of Marketing, Vol. 62, January, pp. 33-47.
Oliver, R. (1993), “Cognitive, affective and attribute bases of the satisfaction response”, Journal of
Consumer Research, Vol. 20, pp. 418-30.
Opiela, N. (2000), “Client online trading: opportunity or obstacle?”, Journal of Financial Planning,
July, pp. 54-60.
Parasuraman, A., Zeithaml, V.A. and Berry, L.L. (1985), “A conceptual model of service quality
and its implications for future research”, Journal of Marketing, Vol. 49, Fall, pp. 41-50.
Parasuraman, A., Zeithaml, V.A. and Berry, L.L. (1988), “SERVQUAL: a multiple-item scale for
measuring consumer perceptions of service quality”, Journal of Retailing, Vol. 64 No. 1,
pp. 12-40.
Paulin, M. and Perrien, J. (1996), “Measurement of service quality: the effect of contextuality”, in
Kunst, P. and Lemmink, J. (Eds), Managing Service Quality, 3rd ed., Chapman, London,
pp. 257-73.
Perreault, W.D. and Leigh, L.E. (1989), “Reliability of nominal data based on qualitative
judgments”, Journal of Marketing Research, Vol. 26, May, pp. 135-48.
Pitt, L., Berthon, P. and Watson, R. (1999), “Cyberservice: taming service marketing problems
with the World Wide Web”, Business Horizons, January/February, pp. 11-18.
Sasser, W.E. Jr, Olsen, R.P. and Wyckoff, D.D. (1978), Management of Service Operations: Text
and Cases, Allyn & Bacon, Boston, MA.
Scullion, M. and Nicholas, D. (2001), “The impact of the Web on the stockbroking industry: big
bang 2”, Aslib Proceedings, Vol. 53 No. 1, pp. 3-22.
Stockdale, R. and Standing, C. (2002), “A framework for the selection of electronic marketplaces:
a content analysis approach”, Internet Research: Electronic Networking Applications and
Policy, Vol. 12 No. 3, pp. 221-34.
Unsal, F. and Movassaghi, H. (2001), “Impact of Internet on financial services industry: a case
study of on-line investing”, Managerial Finance, Vol. 27 No. 7, pp. 54-65.
Van Dyke, T.P., Kappelman, L.A. and Prybutok, V.R. (1997), “Measuring information systems
service quality: concerns on the use of the SERVQUAL questionnaire”, MIS Quarterly,
Vol. 21 No. 2, pp. 195-208.
IJSIM Van Riel, A.C.R., Liljander, V. and Jurriens, P. (2001), “Exploring consumer evaluations of
e-services: a portal site”, International Journal of Service Industry Management, Vol. 12
15,3 No. 4, pp. 359-77.
Wazienski, R.J. (2000), “The ethnograph”, Social Science Computer Review, Vol. 18 No. 3,
pp. 351-6.
Wolfinbarger, M.F. and Gilly, M.C. (2002), “.comQ: dimensionalizing, measuring and predicting
326 quality of the e-tailing experience”, working paper, Marketing Science Institute,
Cambridge, MA, pp. 1-51.
Yang, Z. and Jun, M. (2002), “Consumer perception of e-service quality: from Internet purchaser
and non-purchaser perspectives”, Journal of Business Strategies, Vol. 19 No. 1, pp. 19-41.
Yang, Z. and Peterson, R.T. (2002), “The quality dimensions of Internet retail food purchasing: a
content analysis of consumer compliments and complaints”, Journal of Foodservice
Business Research, Vol. 5 No. 2, pp. 25-46.
Yang, Z. and Peterson, R.T. (2003), “I read about it online . . . Web-based product reviews provide
a wealth of information for marketers”, Marketing Research, Vol. 15 No. 4, pp. 26-31.
Yang, Z., Peterson, R.T. and Huang, L. (2001), “Taking the pulse of Internet pharmacies: online
consumers speak out on pharmacy services”, Marketing Health Services, Vol. 21, Summer,
pp. 4-10.
Yoo, B. and Donthu, N. (2001), “Developing a scale to measure the perceived quality of Internet
shopping sites (sitequal)”, Quarterly Journal of Electronic Commerce, Vol. 2 No. 1, pp. 31-47.
Zeithaml, V.A. (2002), “Service excellence in electronic channels”, Managing Service Quality,
Vol. 12 No. 3, pp. 135-8.
Zeithaml, V.A. and Bitner, M.J. (2003), Services Marketing Integrating Customer Focus Across the
Firm, 3rd ed., Irwin/McGraw-Hill, Boston, MA.
Zeithaml, V.A., Berry, L.L. and Parasuraman, A. (1996), “The behavioral consequences of service
quality”, Journal of Marketing, Vol. 60 No. 2, pp. 31-46.
Zeithaml, V.A., Parasuraman, A. and Malhotra, A. (2001), “A conceptual framework for
understanding e-service quality: implications for future research and managerial practice”,
MSI Working Paper Series, No. 00-115, Cambridge, MA, pp. 1-49.
Zeithaml, V.A., Parasuraman, A. and Malhotra, A. (2002), “Service quality delivery through Web
sites: a critical review of extant knowledge”, Journal of the Academic of Marketing Science,
Vol. 30 No. 4, pp. 362-75.