You are on page 1of 14

Journal of Retailing and Consumer Services 76 (2024) 103600

Contents lists available at ScienceDirect

Journal of Retailing and Consumer Services


journal homepage: www.elsevier.com/locate/jretconser

Can chatbot customer service match human service agents on customer


satisfaction? An investigation in the role of trust
Dongling Huang a, 1, Dmitri G. Markovitch b, 1, Rusty A. Stough a, b, 1, *
a
Marketing at the David Nazarian College of Business and Economics, California State University, Northridge, 18111 Nordhoff Street, Northridge, CA 91330-8377, USA
b
Marketing at Maine Business School, University of Maine, Orono, ME 04469, USA

A R T I C L E I N F O A B S T R A C T

Handling Editor: Prof. H. Timmermans Consumers prefer human service over chatbots in some contexts. We draw on treatments of trust in information
systems research to propose a three-tiered model of trust which posits that customer outcomes following chatbot
Keywords: service are influenced by dispositional trust in service technologies, generalized learned trust in chatbots, and
Customer service particularized situational trust in one’s chatbot service-giver. Our empirical tests, based on 643 participants in 3
Chatbots
studies, show that dispositional trust partially mediates the effect of learned trust. We also evaluate two alter­
Artificial intelligence
native chatbot service configurations to help improve customer acceptance. The approach which combined
Trust
Dispositional trust chatbot service with a light touch human intervention matched the human service condition on most customer
Trust propensity outcomes.
Learned trust
Situational trust
Satisfaction
Loyalty
Recommendation acceptance

In an era dominated by technology, humans will always win on perform complex service tasks while using chatbots predominantly in
experience. That’s why Internova Travel Group, one of the industry’s supporting roles. The advent of generative AI programs, such as
largest travel services companies … announced the launch of a new ChatGPT and Bard, promises to finally enable full automation of
advertising campaign to remind consumers that choosing to book travel customer service encounters.
through humans may be the best travel decision they could ever make Chatbot-based customer service offers critical advantages over
(Internova Travel Group, 2021). human service, such as lower cost at scale, 24/7 on-demand availability,
Along with ease of use, trust is a key component of a consumer’s and scalability (i.e. Pizzi et al., 2021). A customer service chatbot
satisfaction with a digital experience (Gartner, Inc. 2018). platform adopted in a single organization can be made uniform with
respect to agent training and expertise, and be unaffected by factors that
1. Introduction normally degrade human performance, such as fatigue, memory lapses,
and distractions. Understandably, organizations have a powerful
A chatbot is a computer program designed to conduct human con­ incentive to replace human agents with artificial intelligence (AI)
versation via text online or via voice using text-to-speech (Ciechanowski chatbots in customer service (Wang et al., 2022; Sands et al., 2021).
et al., 2019; Mariani et al., 2023). Although chatbots have existed since To achieve success with full automation, organizations will first need
the 1960s, dramatic advances in artificial intelligence (AI) and natural to overcome consumer mistrust of chatbot service. Surveys of consumer
language processing are driving a recent wave of chatbot adoption in sentiment toward AI consistently show that consumers view chatbots as
commercial customer service applications (i.e., Aleedy et al., 2019; being less understanding, empathetic or flexible, and harder to deal with
Shumanov and Johnson, 2021). Yet, until recently, chatbots’ functional than human agents (e.g., Li et al., 2023; Hsu and Lin, 2023; Clark, 2022;
limitations have pressed organizations to rely on human agents to Gartner, 2018; Sheehan et al., 2020; Beaner, 2022; Pegasystems, 2019;

* Corresponding author. Maine Business School. 5723 DP Corbett Business Building, Orono, ME 04469.
E-mail addresses: dongling.huang@csun.edu (D. Huang), dmitri.markovitch@maine.edu (D.G. Markovitch), rusty.stough@maine.edu (R.A. Stough).
1
The authors contributed equally to this research. The names are listed alphabetically.

https://doi.org/10.1016/j.jretconser.2023.103600
Received 26 August 2023; Received in revised form 7 October 2023; Accepted 9 October 2023
Available online 17 October 2023
0969-6989/© 2023 Elsevier Ltd. All rights reserved.
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

Rese et al., 2020; Hill et al., 2015). The referenced studies all report tested extensively before launch (Bowman, 2023; Roose, 2023). A key
lower consumer trust in and preference for chatbot service, compared challenge is that such novel behaviors are hard to predict and correct
with human agents. Of note, the most capable generative AI chatbots because they are dynamic (e.g., see Bowman (2023) for an overview and
making the headlines are controversial for different reasons, and raise a Simon Willison’s weblog at https://simonwillison.net/2023/for current
host of other concerns involving their ethics, accuracy, and trustwor­ commentary on generative AI).
thiness (e.g., Bowman, 2023; Pegoraro et al., 2023; Roose, 2023; Jabeur Importantly, trust is known to affect perceptions of customer service
et al., 2023). provision, value, and outcomes as experienced or perceived. Trust in­
Inspired by the practical challenge of achieving consumer accep­ fluences how consumers experience exchanges and interactions by
tance of chatbot service, we set out to explore the role of trust in how reducing perceived risk, anxiety, and discomfort (McLean et al., 2021;
consumers experience and respond to chatbots. We propose a novel Pavlou, 2003). Thus, higher trust in a provider can make service in­
three-tiered model of trust in chatbots which posits that customer out­ teractions relatively more pleasant and satisfying. Trust motivates
comes following chatbot service provision and, hence, chatbot accep­ greater acceptance of provider recommendations, which, depending on
tance, are influenced by one’s propensity to trust service technologies, the service context, may enable better regimen compliance or more
generalized learned trust in chatbots, and particularized situational trust appropriate product selection and use (McKnight et al., 2002), thus
in one’s chatbot service-giver. This is our first contribution. Our second, enabling better results and user experience. Consumers who trust the
empirical, contribution consists in testing the impact of trust on con­ provider also show greater commitment and belief in the value of its
sumer chatbot acceptance. This contribution addresses a notable gap in offerings while spending less effort to consider alternative solutions
the literature (see Web Appendix A for a summary). Despite the prolif­ (Morgan and Hunt, 1994). Consistent with these results, empirical
eration of theoretical and empirical articles addressing the impact of AI research finds a positive relationship between trust and customer
on marketing phenomena, empirical research exploring the impact of satisfaction, including as a mediator variable, although the direction of
trust on customer experience with chatbot service has been quite causality has not been fully established (Chu et al., 2012; Geyskens et al.
limited. 1998).
We present three studies, collected online in the United States of
America, which test our research hypotheses. We use Study 1 (n = 199) 2.2. The trust construct
to confirm that individuals will report higher customer satisfaction,
repatronage intentions, and service agent’s recommendation acceptance Most researchers conceptualize trust based on their specific disci­
(henceforth referred to as our focal customer outcomes) when served by plinary perspective, which has given rise to varying definitions of trust
a human agent versus a chatbot in our research setting. Study 1 also (McKnight et al., 2002). In customer service situations, customers can
explores whether an individual’s (1) propensity to trust service tech­ never be certain that the provider’s representative is knowledgeable and
nologies or (2) generalized learned trust in chatbots impact the focal genuinely willing to help. Hence, the marketing literature tends to
customer outcomes when individuals are served by a chatbot. Study 2 (n emphasize in its conception of trust the trustor’s confidence in the
= 225) primes trust in the chatbot service provider to test whether trustee’s competence and reliability. For example, Morgan and Hunt
manipulating provider-specific situational trust can impact customer (1994, p. 23) define trust as the feeling of “confidence in the exchange
experience with chatbot service. Study 3 (n = 219) evaluates two partner’s reliability and integrity.” Garbarino and Johnson (1999, p. 73)
chatbot service configurations informed by our trust framework to see if define trust in organizational settings as “customer confidence in the
they can match the human service condition on the focal customer quality and reliability of the services offered by the organization.”
outcomes. Barber (1983, p. 14) notes that trust generally entails “the expectation of
a technically competent role performance.”
2. Theory and hypotheses Various models of trust have been proposed in psychology, sociol­
ogy, engineering, information systems, and the business disciplines.
2.1. The role of trust in consumer acceptance of chatbot customer service Most models identify the trustor (e.g., a consumer), the context, and the
trustee (e.g., a provider’s sales or service agent) as distinct sources of
Trust plays a central role in business and social transactions and variability in trust (Lee and See, 2004; Mayer et al., 1995). The mar­
relationships (Dwyer et al., 1987). It influences consumers’ willingness keting literature, including service research, has, most extensively,
to engage in market exchanges and rely on providers (Garbarino and employed theoretical frameworks and measures centered on the trustee.
Johnson, 1999). Trust is a known factor in automation acceptance (Hoff This literature emphasizes the critical importance of the provider’s
and Bashir, 2015). Research in the “computers as social actors” para­ perceived competence, integrity, and benevolence in instilling trust (e.
digm shows that people react socially to computers, which also places g., Basso and Pizzutti, 2016; Doney and Cannon, 1997; also, see a
trust at the center of human-computer interactions (e.g., Nass and Moon, meta-analysis by Geyskens et al. 1998).
2000; Xu et al., 2022). Glikson and Woolley (2020, p. 230) note that Our theoretical development deviates from this paradigm in that it
“trust is particularly relevant to the human-AI relationship because of builds on trust frameworks developed in the information systems liter­
the perceived risk embedded in human-AI relations, due to the ature and human factors research, which are both thematically related
complexity and non-determinism of AI behaviors.” In other words, to our exploration. The specific model we build on argues that the
consumers may never sufficiently understand AI and, hence, are trustor-, trustee-, and context-based factors find a distinct expression in
compelled to rely on some degree of trust when interacting with AI three inter-related trust layers: dispositional trust, learned trust, and
agents. situational trust (Hoff and Bashir, 2015; Mariani et al., 2023; Marsh and
Trust is particularly important in consumer acceptance of generative Dibben, 2003; McKnight et al., 2002). Dispositional trust encompasses a
AI chatbots as service-givers. These chatbots combine highly evolved trustor’s stable personality characteristic of being more or less trusting
capabilities that enable human-level performance with fundamental towards others. Learned trust reflects one’s trust behavior towards a
limitations that undermine their trustworthiness. Central to the particular entity arising from experience and other learning behaviors,
contradiction is the fact that generative AI uses machine learning to infer whereas situational trust reflects one’s trusting stance adjusted based on
information. This process is known to produce inaccuracies and outright situational cues present during an exchange (e.g.,Marsh and Dibben,
falsehoods, which, nevertheless, sound highly plausible. Equally 2003). Hence, the model presents a dynamic perspective on trust which
important, “as language models (LMs) scale, they develop many novel evolves in the moment and over time.
behaviors, good and bad” (Perez et al., 2022, p. 1). For example, the Nevertheless, the prevalent approaches to studying trust in tech­
early version of Bing Chat displayed troubling behavior in spite of being nologies show a notable limitation. As McKnight et al. (2011, p. 2) point

2
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

out, even “within the Information Systems (IS) domain, as in other technology (Rau et al., 2009). However, longer-term, dispositional trust
fields, trust is usually examined and defined in terms of trust in people may evolve, influenced by accumulating experiences (Hoff and Bashir,
without regard for trust in the technology itself.” Therefore, we extend 2015), which initially inform learned trust. People also develop greater
the model to reflect the specificities of service interactions involving a trust towards other individuals and technologies as they age (e.g.,
non-human trustee. Our model proposes three layers of trust which, a Greiner and Zednik, 2019). For example, Ho et al. (2005) find that older
priori, are relevant to chatbot acceptance—(1) dispositional trust to­ individuals trust and rely more on technology-based decision aids than
wards service technologies, (2) learned trust in chatbots, in general, and younger people.
(3) situational trust in the specific chatbot providing a service. Of note, Importantly, the degree of one’s trust in a trustee directly influences
this distinct combination of trust factors reflects two ubiquitous service one’s experiences with the trustee, including satisfaction (Chu et al.,
scenarios—when a consumer interacts with a particular service pro­ 2012; Geyskens et al. 1998). “Trusting stance means that one assumes …
vider’s chatbot for the first time or only once. In either case, there is no better outcomes from dealing with people as though they are well
particularized learned trust in the chatbot representing the provider, but meaning and reliable” (McKnight et al., 2002, p. 340). These trust effects
all consumers will vary in terms of dispositional trust, and have prior also extend to human-AI interactions (for an overview of the literature
general experience involving chatbots or AI assistants and, hence, have on human reactions to AI, see Glikson and Woolley, 2020). Hence, in­
developed more or less generalized learned trust in chatbots. dividuals with higher dispositional trust in technology are likely to
perceive their chatbot interactions in a more favorable light than in­
2.3. Trust in people versus trust in technology dividuals who score low on dispositional trust in technology. Formally.
H1. Dispositional trust in service technologies is positively associated
There are important conceptual similarities between trust in people
with customer satisfaction, repatronage intentions, and recommenda­
and trust in technology. Both trust phenomena arise when one chooses
tion acceptance following interactions with chatbot customer service.
to rely on another party (i.e., either a human or a chatbot) to achieve an
objective under uncertainty and lack of total control. Both phenomena
2.5. Learned trust
also involve expectations of competent performance, although, a chat­
bot’s perceived competency will derive from its perceived adequacy for
The development of trust is a dynamic process (Lee and See, 2004).
the task (van Wezel et al., 2021). In both cases, there is an expectation of
In the absence of relevant information or prior experience, individuals
benevolence—that the agent’s purpose is to help, and also that human or
may adopt a trusting stance towards a service provider based on their
chatbot service agents will be honest with the customer (Chung et al.,
trust propensity. Subsequently, consumer trust towards a service pro­
2020). Hence, chatbots can be evaluated based on the same fundamental
vider will evolve as consumers interact with the provider, experience
trust antecedents as human trustees. Moreover, the nomological
service outcomes, and update their knowledge structures, beliefs, and
network surrounding trust in technology is similar to trust in people and
attitudes (Mayer et al., 1995; McKnight et al., 2002). Hence, personal
includes the common trust types and linkages attributed to interpersonal
experience is a key mechanism in trust development, although, this trust
relationships (McKnight et al., 2011).
layer is also influenced by word of mouth (e.g., Meuter et al., 2013; Kong
et al., 2020), and various learning processes, including observational
2.4. Dispositional trust in service technologies
learning (Zhang, 2010). Accordingly, the resultant form of trust is often
referred to as learned trust. Where it exists, learned trust is believed to
Mayer et al., 1995, p. 715), in their widely cited work, define
supersede dispositional trust in its impacts on consumer behavior
dispositional trust as an individual’s “general willingness to trust
(Marsh and Dibben, 2003).
others.” Also called trust propensity, dispositional trust represents a
Learned trust initially develops in the form of particularized trust, or
personality trait that reflects an individual’s tendency to see the best in
trust towards a specific trustee (e.g., Glanville and Shi, 2020). Over time,
others and have a kind of belief in humanity that leads one to be more
as individuals interact with (or learn about) multiple entities in a subject
trusting irrespective of other factors, including the provider’s attributes
class, they develop trusting beliefs towards the entire class of similar
(McKnight et al., 1998).
trustees. An accumulation of positive experiences, hence, engenders
In a similar vein to interpersonal trust, individuals display variability
generalized learned trust towards the ‘average’ representative of the
in trust propensity towards technology in general (Hoff and Bashir,
class (e.g., Freitag and Traunmüller, 2009). Various studies in the field
2015), as well as distinct classes of technologies (McKnight et al., 2011).
of robotics, automation, and AI find a similar trajectory in human trust
In the latter case, a trusting stance towards a class of technologies rep­
towards AI (Glikson and Woolley, 2020; Ullman and Malle, 2017). For
resents a lower-order construct, whereby trusting beliefs emanate from
instance, Bartneck et al. (2009) show that a positive interaction with a
one’s dispositional trust in technology (Serva et al., 2005). This pre­
robotic pet improved the study participants’ attitude towards robots in
disposition to trust specific technologies essentially reflects beliefs that
general.
the technologies have beneficial attributes. For example, an individual
Based on this literature, we assume that learned trust in some service
with a higher propensity to trust in service technologies (e.g., automated
providers extends to similar provider types. In particular, a consumer
teller machines (ATMs), self-service kiosks, automated customer service,
who has experienced satisfactory chatbot-based customer service by
etc.) is more likely to assume that these technologies are reliable,
some providers is more likely to develop generalized trust in chatbots
functional, safe, useful, and help to improve one’s quality of life
and expect similarly satisfactory outcomes in future service encounters
(McKnight et al., 2011). The trust-in-automation research stream fea­
involving other chatbots. Drawing on the research which shows a pos­
tures several studies which explore dispositional trust in particular
itive relationship between trust and one’s experiences with a trustee
technologies. For instance, Biros et al. (2003) report that individuals
(McKnight et al., 2002), we posit that higher learned trust in chatbot
with a higher dispositional trust in computers showed greater accep­
service, in general, will favorably impact customer outcomes following
tance of inputs from drones.
specific chatbot service encounters. Stated formally:
Unlike learned trust, which evolves following every interaction, and
situational trust, which is influenced by context in the moment, dispo­ H2. Generalized learned trust in chatbots is positively associated with
sitional trust tends to be relatively stable across situations, because it is a service satisfaction, repatronage intentions, and recommendation
feature of one’s psychological makeup. People with different life’s ex­ acceptance following interactions with chatbot customer service.
periences, personality traits and cultural backgrounds differ in their Moreover, scholars often model trust as a mediating variable which
dispositional trust. For example, the Japanese are known to be more links various antecedents with important outcomes, including those
trusting than others (e.g., Fukuyama, 1995), including towards robotic studied in this research (for an overview, see meta-analysis by Geyskens

3
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

et al. 1998). For example, Benamati et al. (2009) report that trust in an trustee or the service environment (e.g., Geyskens et al. 1998). Unlike
online vendor mediates consumer intentions to transact with its website. dispositional trust, which represents a durable personality trait, internal
In their study of supplier-retailer dyads, Morgan and Hunt (1994) find situational trust antecedents are more transitory and context-specific.
support for trust as a key mediator between relationship antecedents, One widely explored trust antecedent is domain expertise, or con­
such as expected benefits and shared values, and relationship outcomes, sumer knowledge (Hoff and Bashir, 2015). Expertise has a profound
including acquiescence, loyalty, and cooperation. Auh (2005) extends impact on consumer behavior (Alba and Hutchinson, 1987), including
Morgan and Hunt’s key mediating variable model to show that trust the extent of information search and acceptance of external product
fully mediates the effect on loyalty of a service provider’s social and reviews. For example, Chinchanachokchai et al. (2021) show that
relational (i.e., soft) attributes, but only partially mediates the effect of domain experts have higher confidence in user-generated product rec­
hard attributes, such as perceived performance quality. In research that ommendations than AI-generated recommendations. Other internal
is directly related to our investigation, Lin et al. (2021) examine con­ factors influencing situational trust include the trustor’s self-efficacy,
sumer acceptance of a virtual salesperson’s recommendation when it mood, and attention capacity during an interaction (Hoff and Bashir,
conflicts with other customers’ recommendations. They find that trust in 2015; Lee and See, 2004). For example, greater computer self-efficacy
a virtual salesperson fully mediates the link between conflicting reviews positively impacts trust in decision support systems (DSS), compli­
and the willingness to follow a virtual agent’s recommendation. In ance, and DSS adoption (Madhavan and Phillips, 2010).
contrast, Ameen et al. (2021) report that consumer trust only partially External influences on situational trust are contextual cues that
mediates the perceived impact of AI-enabled service quality, conve­ impact consumer perceptions of a service provider’s trustworthiness.
nience, and personalization on AI-enabled customer experience. Hence, some of these factors pertain to the task and circumstances,
Consistent with this, the trust literature points to a likely causal including task difficulty, task framing, and risk perceptions (Hoff and
ordering between dispositional and learned trust. Dispositional trust is Bashir, 2015). Structural characteristics of the service environment also
typically treated as the highest-order construct. It is believed to influ­ play a notable role (McKnight et al., 2002. These are third-party
ence one’s behavior in risky unfamiliar contexts, including interactions mechanisms that motivate situational trust by minimizing the
with unfamiliar (types of) service providers and AI chatbots (Hasan perceived risk or otherwise enhancing the likelihood of the provider
et al., 2021; Schultz, 2006). Over time, the other party’s trustworthiness performing as expected. Examples of such “structural assurance” factors
is revealed. In this way, dispositional trust acts as an important mech­ include legal protections, protections from credit card companies, es­
anism which facilitates experiential learning and learned trust forma­ crows, warranties, third party certifications, and seals of approval
tion. Subsequently, in situations involving familiar trustees, one’s (Burke, 2002; Kennedy et al., 2001).
behavior and outcomes will be influenced increasingly by learned trust. In addition, a service provider’s attributes may greatly influence
The impact of dispositional trust on outcomes will also increasingly flow perceptions of trustworthiness (e.g., Chen and Dhillon, 2003; Doney and
through learned trust. In other words, learned trust represents a Cannon, 1997). Several authors propose models which present con­
second-order construct which likely mediates the effect of dispositional sumer trust as a function of perceived provider competence, benevo­
trust. Formally: lence, integrity, and reliability (Mayer et al., 1995; McKnight et al.,
2002). Consumer perceptions of these attributes may develop over
H3a. Generalized learned trust in chatbots mediates the effect of
multiple interactions with the provider, and, hence, motivate learned
dispositional trust in service technologies on satisfaction, repatronage in­
trust, as detailed in the preceding section. Or, they may be informed by
tentions, and recommendation acceptance following interactions with
new contextual cues and impact consumer trust during the encounter.
chatbot customer service.
The latter may include the provider’s apparent professionalism (Gregori
Our hypothesis H3 is motivated by the view of causal relationships
et al., 2014), brand recognition (Doney and Cannon, 1997), evidence of
between trust components which is prevalent in the trust literature (e.g.,
offline presence (Chen and Dhillon, 2003), or website characteristics,
Mayer et al., 1995; McKnight et al., 2011). However, alternative per­
such as ease of use, usefulness, or even visible links to human customer
spectives also exist. There is a less well-developed, primarily theoretical,
support (Gefen, 2000). Multiple authors also find that provider
view which points to dispositional trust as a possible mediator of learned
communication, reputation, and evidence of past performance are
trust. First, there is evidence of reverse causality. One’s trust propensity
particularly germane to trust formation (e.g., Chen and Dhillon, 2003;
is relatively stable, but it is not fixed—changes simply occur over a
Geyskens et al. 1998). Importantly, the same provider attributes that
longer time-frame. They are influenced by accumulated experiences and
influence consumer trust in human agents also impact trust in AI tech­
feedback concerning outcomes linked to trusting behaviors (Hoff and
nologies (Glikson and Woolley, 2020). In addition, technology-related
Bashir, 2015; McKnight et al., 2011). Additionally, technologies evolve,
attributes, such as disclosure of information about the operating rules
and dispositional trust will reflect updated beliefs that capture tech­
governing the technology, may further affect situational trust in AI
nology advancement and learned trust experience. In other words,
(Glikson and Woolley, 2020; Hoff and Bashir, 2015).
learned trust also induces changes in dispositional trust. Moreover,
In sum, extant research indicates that presenting favorable infor­
dispositional trust may continue to affect one’s trusting beliefs and be­
mation about a service provider’s attributes, certifications, success rate,
haviors after trustworthiness has been established. It acts as a lens
and other measures of past performance may have a priming effect and
through which one sees social interactions and which is used to interpret
enhance consumers’ particularized situational trust in that provider.
trustees’ actions (Govier, 1994; Lewis and Weigert, 1985). We formalize
Higher trust, in turn, should result in better consumer evaluations of
this perspective on the role of dispositional trust as a competing hy­
their service experiences with the provider. Hence, we test the following
pothesis as follows:
hypothesis:
H3b. Dispositional trust in service technologies mediates the effect of
H4. Individuals with higher particularized situational trust in a chatbot
generalized learned trust in chatbots on satisfaction, repatronage in­
providing service will display higher satisfaction, repatronage in­
tentions, and recommendation acceptance following interactions with
tentions, and recommendation acceptance following their service in­
chatbot customer service.
teractions with the chatbot than individuals with lower particularized
situational trust.
2.6. Situational trust
3. Manipulating trust to improve customer outcomes
Situational trust is the trust that develops during a particular service
interaction in a particular setting. It is a function of multiple factors,
Since improving customer trust outcomes in chatbot service
some of which are internal to the trustor, and others—specific to the

4
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

encounters is an important motivation for this research, we next element with a lower-trust element in a system should additively in­
consider the performance of two practical service configuration ap­ crease the total trust in the system (e.g., Amin-Naseri and Gilbert, 2014).
proaches suggested by our theoretical framework. The customer out­ The established positive relationship between trust and satisfaction (e.
comes attained by a human service agent represent a natural benchmark g., Chu et al., 2012) means that consumers will likely feel more positive
for this investigation. about the total service experience delivered by a chatbot-human agent
The preceding discussion indicates that situational trust towards a team.
specific provider can be influenced by contextual cues. This malleability The second mechanism whereby a human agent’s participation
makes situational trust an ideal focal point for a marketing action. As should boost trust in the chatbot component of the service team involves
hypothesized in H4, higher particularized situational trust is expected to the well-documented halo effect. It is a cognitive bias which entails the
be associated with generally better customer outcomes. This enables us tendency for positive impressions of a salient person or characteristic to
to set up our first “non-inferiority” performance test which we express in influence how other related persons or characteristics are judged (e.g.,
the following hypothesis:2 Nisbett and Wilson, 1977). The halo effect has been observed in a variety
of team and system contexts, which are, essentially, similar to our pro­
H5. Following a service encounter, individuals with high particularized
posed “service team” or “service system” setting. For example, business
situational trust in their chatbot service-giver will report lower satisfac­
scholars find that managers tend to provide more positive evaluations of
tion, repatronage intentions, and recommendation acceptance than in­
subordinates who work in close proximity to them (Ferris et al., 1994).
dividuals in the human-service-agent condition.
In systems research, studies routinely find that multicomponent ratings
exhibit higher than expected correlation among the rated components.
3.1. Low-cost combination of chatbot and human customer service For example, movie ratings display an elevated correlation across the
typical movie attributes (Sahoo et al., 2012). Moreover, the halo effect
While fully automated customer service may be the ultimate goal for extends to perceived trustworthiness of technologies. For example,
most organizations, consumer mistrust of chatbot service in some con­ Gabrieli et al. (2021) report that placing images of trustworthy-looking
texts may motivate a gradual transition whereby chatbots will be com­ human faces on ATMs increased the perceived trustworthiness of the
bined with human agents in distinct ways to achieve operational or ATMs compared with the ATMs that featured less trustworthy-looking
marketing objectives.3 This projection is supported by practitioner sur­ faces.
veys which show that a higher proportion of consumers still prefer to be Based on these mechanisms, we proceed to evaluate the proposed
served by humans rather than chatbots even for basic tasks, such as chatbot-human service configuration on the focal performance measures
opening an account or making a purchase (e.g., Beaner, 2022; Chen relative to the human-only configuration. We are primarily interested in
et al., 2022). To the extent that the mistrust in chatbot service may be the chatbot-human configuration’s performance under conditions of low
deep-rooted and “sticky,” only manipulating particularized situational situational trust in the chatbot. However, it is also relevant to consider
trust may be insufficient to achieve satisfactory customer outcomes. We, the impact that adding a human agent may have when consumers have
hence, consider whether using a chatbot in the lead role with human high trust in the chatbot. We use the non-inferiority experimental design
support may engender higher consumer trust overall. to test the following set of hypotheses:
One practical chatbot-human service configuration that seemingly H6a. The proposed chatbot-human service configuration will under­
satisfies efficiency objectives could feature a front-line chatbot agent perform human-only service on the focal customer outcomes under
augmented with a “light-touch” human intervention at the conclusion of conditions of low particularized situational trust in the service chatbot.
a service encounter. In this setup, the chatbot would be tasked with
resolving a customer issue or making a sale in its entirety. In situations H6b. The proposed chatbot-human service configuration will under­
flagged as non-routine, the encounter would conclude by handing the perform human-only service on the focal customer outcomes under
consumer over to a human service specialist whose objective would be conditions of high particularized situational trust in the service chatbot.
to reassure the consumer and upsell, where appropriate. The human We summarize our hypotheses and key flows in Fig. 1. Note that the
intervention would be selective, last seconds in many cases, and, consumer preference factors across subjects are controlled through
therefore, cost little per service encounter on average while generating random assignment to our experimental conditions. The service quality
incremental sales revenue. element of the model is a major factor which affects customer outcomes.
There are two plausible mechanisms whereby even a fleeting human It is assumed to be constant across all our conditions.
agent’s addition to chatbot-led service may benefit customer outcomes.
First, consumers currently perceive human service agents as being 4. Research method
generally more trustworthy than chatbots (e.g., Beaner, 2022; Clark,
2022; Gartner, Inc. 2018; Pegasystems, 2019). Based on our framework, 4.1. Participants and Procedure
this engages higher dispositional and learned trust, since these trust
layers are interlinked through a feedback loop. Combining a higher-trust We conducted three experimental studies using the vignette method
(e.g., Jörling et al., 2019; Harborth and Pape, 2021) to test our research
hypotheses. We used Amazon’s MTurk platform to recruit study par­
2
Non-inferiority experimental designs, which are prevalent in medicine, test ticipants and recruited participants from the United States in the fall of
whether a proposed treatment is at least as good as the standard-of-care. A 2022. We used a power analysis to ensure adequate samples. Partici­
positive result can help establish that switching to the new treatment will do no pants were randomly assigned to six experimental conditions across the
harm. Our experimental challenge is essentially the same—we seek to identify three studies. All the studies opened with scales measuring (1) gener­
service configurations that will at least match human service on key customer alized learned trust in chatbots and, separately, learned trust in human
outcomes. Since demonstrating non-inferiority relative to the human service service agents, and (2) dispositional trust in service technologies and,
benchmarks is the objective, we adopt the “non-inferiority” testing framework.
separately, dispositional trust in service personnel. We randomized the
This entails framing the null hypotheses so that the new treatment is posited to
order in which the question blocks about chatbots and human agents
be inferior to the benchmark. The alternative hypothesis is that the new
treatment is non-inferior. A non-inferiority claim is supported if the null hy­ were presented.
pothesis (of lower performance) is rejected. Next, we introduced our experimental vignette (see Web Appendix B
3
Although organizational adoption of generative AI is not the focus of this for vignettes and primes) which had the same opening paragraph in all
research, it is noteworthy that business leaders also cite trust as a barrier to AI three studies. It instructed the participants to imagine that they were
adoption, according to a recent IBM survey (Wiggers, 2023). reserving a long-planned African safari, accompanied by a renowned

5
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

Fig. 1. A model of the three layers of trust and customer service outcomes. The dashed arrow shows the hypothesized alternative mediation pathway. Feedback
effects are not shown for parsimony reasons. HM designates human-agent service. Hypotheses in bold font have received empirical support. * Hypotheses H5, H6a,
and H6b were stated as inferiority hypotheses. Only H6b received non-inferiority support. See Footnote 1 for details.

tour guide. Participants imagined having an initial interaction online (Stella)-human agent service combination with participants being
with a reservation agent who was, depending on the condition, either assigned either to high-trust or low-trust conditions. As in Study 2,
human (Amy Sullivan) or a chatbot (introduced as “Stella, your AI Stella’s recommendation of a replacement tour guide included the same
Virtual Representative”). Participants were then shown a mock-up of additional information about Stella’s track record and implications for
their imagined online chat with the reservation agent. In all the studies, the decision. The participants were then instructed to imagine that they
the participants were informed that their much-preferred tour guide was accepted the reservation agent’s recommendation and concluded the
unavailable. Instead, the reservation agent recommended another, reservation details, at which point Stella handed the customer over to
obscure, tour guide, who was new to the company. There was no public the company’s human customer relations manager Marisa Diaz. Marisa
information available about the replacement guide, except for the reassured the customer about the replacement tour guide and offered
reservation agent’s assurance that the guide was “experienced and well additional tour options which were declined. This concluded the reser­
trained, like all our guides.” This introduced an element of uncertainty vation interaction in Study 3.
and risk—the customer was reserving the vacation of a lifetime at a high
cost, but not getting everything they were hoping to get. Moreover, a key 4.2. Screens and attention checks
element of the vacation experience was fundamentally uncertain and, at
a minimum, not as good as the preferred tour guide. From this point, the We took several steps to ensure data quality. We asked all partici­
vignettes varied across the studies in the following manner. pants whether they had ever made a hotel or vacation reservation and
In Study 1, the participants interacted either with Amy Sullivan, a been on a tour (of any length) with a tour guide. We excluded those who
human agent, or Stella, the chatbot. The participants did not receive any had not, since they may have difficulty relating to our research scenario.
further information to facilitate their decision. They were instructed to We also used three comprehension and attention checks and a separate
imagine that they accepted the reservation agent’s guide recommen­ check to ensure that participants served by Stella understood that Stella
dation and concluded the reservation. This study additionally embedded was a chatbot. We excluded participants who failed any of these checks.
filler information in both conditions in order to create visual and length Using these filters and checks reduced our total sample by 89 (out of
equivalency with our experimental conditions in Studies 2 and 3. We 732) participants and resulted in sample size variation across the
used this study to test hypotheses 1–3. conditions.
In Study 2, which tested hypotheses 4 and 5, the participants inter­
acted only with Stella, the chatbot, but were randomly assigned to high-
4.3. Measures
situational-trust or low-trust conditions. Stella’s recommendation of a
replacement tour guide included additional information about Stella’s
All the response alternatives in our rating scales discussed below are
track record and unpacked its implications for the reservation decision.
measured based on a seven-point scale. Table 1 shows the construction
In the high-trust condition, Stella’s track record showed a 95 percent
of our key variables, controls (described in the following sections), and
customer satisfaction rating with Stella’s recommendations. The
the associated reliability coefficients.
vignette further interpreted the information for the participants,
Dependent variables. To test customer satisfaction we first use a two-
restating that Stella almost always made good recommendations. In the
item measure, Satisfaction_Index, which captures (1) the extent to which
low-trust condition, Stella’s track record showed a 50 percent customer
participants were satisfied with the reservation interaction overall and
satisfaction rating with Stella’s recommendations. The vignette further
(2) satisfaction with the reservation agent specifically. We also measure
interpreted the information by stating that Stella made good recom­
perceived service quality satisfaction using the distinct dimensions of
mendations about half of the time. Study 2 participants were then
the SERVPERF scale (Cronin and Taylor, 1994). Over the years, elements
instructed to imagine that they accepted the reservation agent’s
of SERVPERF have been extensively employed to evaluate service
recommendation and concluded the reservation.
quality satisfaction during distinct service encounters (e.g., Elliott,
Study 3, which tested hypotheses 6a and 6b, featured a chatbot
1994; Rodrigues et al., 2011). Following the convention, we adapt

6
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

Table 1 Table 1 (continued )


Variable construction and construct reliability (coefficient alpha). DVs Items Alpha
DVs Items Alpha
3. Involvement How do you feel about African safaris? 0.88
1. Satisfaction index i. How satisfied or dissatisfied were you 0.87 i. 1 Important – 7 Unimportant a
with your reservation interactions overall? ii. 1 Uninterested – 7 Interested
ii. How satisfied or dissatisfied were you iii. 1 Unexciting – 7 Exciting
with your reservation representative? 4. Dispositional trust in i. I believe that people are basically moral. 0.89
SERVPERF Scale I think that the Customer Service … people ii. I avoid contacts with others.a
2. Efficiency i. Is well organized. 0.82 iii. I believe that most people would lie to
ii. Makes it easy for customers to get their get ahead.a
issues addressed. iv. I can get along with most people.
3. Responsiveness i. Provides prompt service. 0.83
Notes: All the items were measured on a 1–7 scale.
ii. Is willing to help. a
4. Empathy i. Has the customers’ best interests at its 0.91 The item was reverse-scored.
b
core. The measure was used in the human-agent condition.
ii. Provides individualized attention.
iii. Understands the needs of their
SERVPERF to reflect our specific context. Our scale has five dimensions:
customers.
5. Reliability i. Provides service as promised. 0.8
perceived service responsiveness, empathy, reliability, assurances, and
ii. Is error free. efficiency. The latter dimension captures the burden on consumers in
6. Assurances i. Instills confidence in its ability to 0.87 order to receive service. It replaces the “tangibles” dimension in
provide service. SERVPERF, which is less relevant to our e-service context. We treat each
ii. Is knowledgeable.
SERVPERF dimension as a separate dependent variable in order to
7. Provider loyalty i. If you were to make a travel reservation
in the future, how likely would you be to capture the focal effects at a more granular level (e.g., Parasuraman
use the same reservation agency/tour et al., 1985).
operator? In addition to our primary customer satisfaction variables, we use
8. Recommendation i. How likely would you be to call again two measures of behavioral intent. We measure loyalty towards the
acceptance and modify your reservation to get your
initially chosen tour guide, Jackson
provider by asking about the likelihood of using the same reservation
Looseyia? a agency on a future travel occasion. This is a single-item measure that
IVs captures a distinct behavioral loyalty dimension—repatronage in­
1. Dispositional trust in i. I believe that most service technologies 0.9 tentions (Jacoby and Chesnut, 1978; Söderlund, 2006). We measure
service technologies are NOT very effective at what they are
recommendation acceptance by asking about the likelihood of calling
designed to do.a
ii. A large majority of service technologies back to modify the reservation so as to get one’s preferred tour guide.
are excellent.
iii. My typical approach is NOT to trust
new service technologies until they prove 4.4. Independent variables and controls
to me that I can rely on them.a
iv. I generally give a service technology the Our measure of dispositional trust in service technologies is based on
benefit of the doubt when I first use it.
Dispositional trust in service i. I believe that most service employees are 0.8
a reduced version of the scale adapted from McKnight et al. (2011).
personnelb NOT very effective in their job roles.a Similarly, we use a five-item scale adapted from Mayer et al., 1995 to
ii. A large majority of service employees construct our measure of generalized learned trust in chatbots. Con­
are excellent. sumers who indicated having interacted with customer service chatbots
iii. My typical approach is NOT to trust
in the previous two years were asked to rate their previous experience
service employees until they prove to me
that I can rely on them.a with the chatbots in terms of the chatbots’ perceived competence,
iv. I generally give service employees the benevolence, and integrity. We entered missing values on the learned
benefit of the doubt when I first interact trust variable where consumers indicated that they had no interactions
with them. with chatbots in the past two years. For the human condition, we use
2. Generalized learned trust On average, the customer service chatbots 0.89
in chatbots …
slightly modified dispositional and learned trust instruments using the
i. were competent and effective same scales that, instead of focus on “service technologies,” reference
ii. performed their role very well “service employees” to capture trust measures that are relevant to the
iii. were honest with me human service condition.
iv. tried hard to help me
Our situational trust in the specific service-giver chatbot (Stella) was
Generalized learned trust in v. acted in my best interests
service personnelb On average, the customer service 0.96 primed rather than measured. The primes are described in the “Partic­
employees … ipants and Procedure” section. We developed the primes through
Items i.-v. as above extensive pretests in which various primes were gauged based on their
Controls ability to instill higher (or lower) perceptions of the chatbot’s compe­
1. Experience w/chatbots i. In the past 2 years, how much experience 0.83
did you have interacting with Virtual
tence, benevolence and integrity. To be practically relevant, our low-
Assistants (i.e. Siri, Alexa, Google trust prime was calibrated to instill a degree of uncertainty about
Assistant)? Stella’s capabilities, but not outright distrust. Our manipulation checks
ii. In the past 2 years, how much in the pretest phase showed a statistically significant separation between
experience did you have interacting with
low trust and high trust conditions on measures of Stella’s trustworthi­
customer service Chatbots online?
2. Experience w/human i. In the past 2 years, how much experience 0.82 ness. We employed a single-item manipulation check which directly asks
agents did you have interacting with human the participants about Stella’s trustworthiness.
customer service representatives online? To account for possible confoundedness in our analyses, we include
ii. In the past 2 years, how much four controls. First, we control for the extent of recent remote in­
experience did you have interacting with
human customer service representatives
teractions with chatbots and AI assistants, and, separately, human
over the phone? customer service. We also control for involvement, since individuals
who feel strongly about a task may behave differently than individuals
who are less involved. We used a modified version of Zaichkowsky’s

7
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

(1985) involvement scale. Participants responded on a 1–7 scale how differences are significant at p < 0.1. We also compared the conditions
important, interesting, and exciting going on an African safari was to based on a separate check about the extent to which the participants
them. The three items were combined in an equally weighted index. were concerned about AI replacing human interaction and jobs, since
Finally, we use a short-form Evans and Revelle (2008) propensity to strong feelings in this regard could bias our results. On average, our
trust people scale to measure dispositional trust in people. Note that our participants displayed low concern about AI, with mean responses
primary interest in this paper is in the trust constructs directly affecting ranging between 3.9 and 4.3 out of seven across the cells.
chatbot perceptions, and we include the dispositional trust in people We also review the pairwise correlations between our trust di­
variable only as a control. mensions. We note a moderately high correlation of 0.53–0.68 between
dispositional and learned trust variables in all our chatbot cells. This is
5. Results consistent with the hypothesized relationship between those trust di­
mensions. In our five chatbot cells, we note a low and mostly insignifi­
As the first step in our analyses, we tested for significant differences cant correlation between those two variables and the dispositional trust in
between experimental conditions on our control and demographic var­ people, which we use as a control. We further note a significant 0.42
iables (see Web Appendix C for descriptive statistics). None of the mean correlation between the more closely related constructs of dispositional

Table 2
Regressions of customer outcome variables on measures of dispositional trust in service technologies and generalized learned trust in chatbot service in the sample that
received chatbot service. These analyses are based on Study 1.
IVs/DVs = − 1 − 2 − 3 − 4 − 5 − 6 − 7 − 8

Sat_Index Serv_Efficiency Serv_Respons. Serv _Empathy Serv_Reliability Serv_Assurances Provider Loyalty Recomm_Accept

Panel A: test of H1
Disp. Trust in Service Tech 0.522*** 0.465*** 0.246** 0.559*** 0.406*** 0.617*** 0.658*** 0.426**
(0.08) (0.09) (0.08) (0.09) (0.09) (0.09) (0.10) (0.13)
Experience w/Chatbots 0.14 0.14 0.13 0.09 0.03 0.09 (0.03) (0.06)
(0.10) (0.12) (0.10) (0.11) (0.11) (0.10) (0.12) (0.16)
Experience w/Human CSRa 0.10 0.01 0.03 0.19 0.00 0.09 0.09 0.04
(0.10) (0.11) (0.09) (0.10) (0.10) (0.10) (0.12) (0.15)
Involvement 0.10 0.10 0.05 0.03 0.12 0.15 0.18 − 0.14
(0.08) (0.09) (0.07) (0.08) (0.08) (0.08) (0.09) (0.12)
Disp. Trust in People 0.01 0.12 0.12 (0.03) 0.06 0.02 0.02 0.02
(0.11) (0.12) (0.11) (0.12) (0.11) (0.11) (0.13) (0.17)
adj. R2 0.348 0.241 0.139 0.338 0.191 0.397 0.309 0.072
F 11.97 7.541 4.323 11.53 5.853 14.58 10.23 2.588
p 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.03
N 104 104 104 104 104 104 104 104
Panel B: test of H2
Learned Trust in Chatbots 0.481*** 0.486*** 0.239** 0.529*** 0.482*** 0.561*** 0.584*** 0.23
(0.09) (0.09) (0.08) (0.09) (0.08) (0.09) (0.11) (0.13)
Experience w/Chatbots 0.17 0.17 0.14 0.13 0.05 0.14 (0.02) (0.07)
(0.11) (0.12) (0.10) (0.11) (0.10) (0.11) (0.13) (0.17)
Experience w/Human CSRa 0.09 0.00 0.03 0.18 0.01 0.08 0.08 0.01
(0.10) (0.11) (0.09) (0.10) (0.09) (0.10) (0.12) (0.16)
Involvement 0.07 0.05 0.04 − 0.01 0.06 0.09 0.14 − 0.16
(0.08) (0.09) (0.08) (0.09) (0.08) (0.08) (0.10) (0.13)
Disp. Trust in People 0.10 0.19 0.18 0.07 0.16 0.15 0.09 − 0.01
(0.12) (0.13) (0.11) (0.12) (0.11) (0.12) (0.15) (0.18)
adj. R2 0.286 0.24 0.123 0.298 0.272 0.329 0.227 0.001
F 8.996 7.32 3.807 9.51 8.465 10.8 6.858 1.026
p 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.41
Nb 101 101 101 101 101 101 101 101
Panel C: Full Model
Disp. Trust in Service Tech 0.365** 0.262* 0.148 0.378** 0.144 0.444*** 0.449** 0.433*
(0.12) (0.13) (0.11) (0.12) (0.11) (0.12) (0.15) (0.19)
Learned Trust in Chatbots 0.239* 0.313* 0.141 0.279* 0.386*** 0.268* 0.287* − 0.0574
(0.11) (0.13) (0.11) (0.12) (0.11) (0.11) (0.14) (0.18)
Experience w/Chatbots 0.146 0.16 0.13 0.108 0.0446 0.111 − 0.0415 − 0.0968
(0.10) (0.12) (0.10) (0.11) (0.10) (0.10) (0.13) (0.16)
Experience w/Human CSRa 0.111 0.0146 0.0404 0.198 0.0216 0.101 0.0975 0.0318
(0.10) (0.11) (0.09) (0.10) (0.09) (0.10) (0.12) (0.15)
Involvement 0.087 0.0661 0.0483 0.00367 0.0704 0.114 0.16 − 0.14
(0.08) (0.09) (0.08) (0.08) (0.08) (0.08) (0.10) (0.13)
Disp. Trust in People 0.0553 0.163 0.163 0.0287 0.143 0.099 0.0399 − 0.0614
(0.11) (0.13) (0.11) (0.12) (0.11) (0.11) (0.14) (0.18)
adj. R2 0.347 0.263 0.13 0.357 0.276 0.413 0.291 0.045
F 9.843 6.96 3.485 10.25 7.368 12.71 7.835 1.794
p 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.11
Nb 101 101 101 101 101 101 101 101

Notes: The regressions included a constant which is not shown.


Standard errors are in parentheses.
*p < 0.05, **p < 0.01, ***p < 0.001.
a
Experience with chatbots/human customer service representatives (CSR) in the previous two years.
b
The samples size (N) is lower in these regressions due to missing values on the Learned Trust in Chatbots variable for individuals who indicated no chatbot exposure
in the previous two years.

8
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

trust in people and dispositional trust in service personnel in the human- paths for dispositional and learned trust. As the first step, we fitted a full
agent cell. Taken together, the observed statistics are consistent with model with both trust variables (shown in Panel C of Table 2). The
our arguments that dispositional trust in people, dispositional trust in resultant main effect coefficients on the dispositional trust variable
service technologies, and generalized learned trust in chatbots represent largely maintained their magnitude and significance, unlike the co­
distinct constructs which capture different dimensions of trust. efficients on the learned trust variable, which declined in magnitude and
We used ordinary least squares regression (OLS) to test our hy­ lost significance in some cases. Next, to test our hypotheses, we ran two
potheses. In addition to the relevant independent variables, all the re­ sets of mediation analyses for each dependent variable (summarized in
gressions included four control variables—(the extent of) experience Table 3). We used Hayes (2018) PROCESS macro with 5000 bootstrap
with chatbots, experience with human customer service, involvement samples. In the first analysis, we specified dispositional trust in service
with respect to African safari vacations, and dispositional trust in technologies as the independent variable and generalized learned trust in
people. chatbots as the mediator. This analysis produced only two significant (at
p < 0.05) coefficients on the indirect effect of dispositional trust, in the
5.1. Study 1 regressions of Service_Efficiency and Service_Empathy. Thus, our data
provide only limited support for H3a. We then evaluated the reverse
The first aim of Study 1 was to establish the baseline preference for mediation pathway with learned trust as the independent variable and
human or chatbot service in our research sample so that we could dispositional trust as the mediator. All but one of the resultant co­
evaluate actions that can make chatbot service more attractive to con­ efficients on the indirect effect were positive and statistically significant
sumers. We fitted the regressions in the combined sample which (at p < 0.05), consistent with H3b. We conclude that, on balance, these
included human-agent and chatbot service conditions with the total of results provide stronger support for H3b than H3a. Additionally, we note
199 observations. The focal variable was the chatbot service-giver that we observe only partial mediation, since the direct path on learned
condition dummy. trust remains significant in six out of our eight regressions (see Panel C of
The coefficients on the dummy variable indicating chatbot service Table 2).
were significantly negative in all the regressions, and the p-values were
below 0.05 in all but one case. Concerning the latter, the coefficient on 5.2. Study 2
the service-giver dummy was not significant in the regression of Serv­
ice_Efficiency (from the SERVPERF construct). While contradictory, this In Study 2, we primed participants for higher or lower situational
result makes intuitive sense, since the chatbots’ defining characteristic trust in the chatbot service-provider. Our purpose was to explore
at this time is their efficiency. Therefore, we conclude that our subjects, whether one’s level of particularized situational trust during a service
on average, strongly prefer human service. encounter impacts customer satisfaction with chatbot service and sub­
We next explored whether one’s propensity to trust service tech­ sequent trusting behaviors. This study used 219 valid observations. As a
nologies and learned trust in chatbots, based on prior interactions, had a manipulation check, we asked participants to rate Stella’s trustworthi­
positive effect on the focal outcomes, as posited in H2 and H3, respec­ ness in making tour guide recommendations. The manipulation check
tively. To test these hypotheses, we ran separate regressions with the means were 5.23 and 3.88 out of seven in the high-trust and low-trust
focal independent variables and controls in the subset served by Stella, conditions, respectively (p < 0.001), indicating that our priming had
the chatbot. We show the regressions results in Panels A and B of successfully created experimental conditions with different levels of
Table 2. We find the effect of dispositional trust in service technologies on situational trust. We next separately regressed our eight dependent
all the dependent variables to be significantly positive at p < 0.01. variables on a dummy variable for situational trust and the controls. The
Our results for the effect of generalized learned trust in chatbots are situational trust variable took the value of one in the high trust condition
largely similar, except in the Reservation_Acceptance regression, which is and zero otherwise. The results are shown in Tables 4 and 5. All the
not statistically significant, implying that dispositional trust may play a regressions are highly significant, with the coefficient on the High_­
relatively more fundamental role in behavioral responses. We explore Situational_Trust dummy being positive and statistically significant (p <
this supposition further in our investigation of mediation paths detailed 0.01 or lower) in all the regressions. We conclude that our data provide
below. We also conclude that, taken together, our findings support H1 strong support for our hypothesis H4.
and H2.
Our hypotheses H3a and H3b specify two alternative mediation

Table 3
Summary of mediation analysis tests based on Hayes (2018) PROCESS macro with 5000 bootstrap samples. These analyses are based on Study 1.
Independent Variable Mediator Dependent Variable b 95% CI Significance

Panel A: tests of H3a


Dispositional Trust Learned Trust Satisfaction_Index 0.129 [-0.050, 0.310] p> 0.05
Dispositional Trust Learned Trust Serv_Efficiency 0.220 [0.049, 0.417]* p< 0.05
Dispositional Trust Learned Trust Serv_Responsiveness 0.101 [-0.059, 0.292] p> 0.05
Dispositional Trust Learned Trust Serv _Empathy 0.191 [0.001, 0.436]* p< 0.05
Dispositional Trust Learned Trust Serv_Reliability 0.264 [-0.014, 0.435] p> 0.05
Dispositional Trust Learned Trust Serv_Assurances 0.191 [-0.021, 0.425] p> 0.05
Dispositional Trust Learned Trust Provider Loyalty 0.174 [-0.004, 0.373] p> 0.05
Dispositional Trust Learned Trust Recomm_Acceptance 0.169 [-0.009, 0.371] p> 0.05
Panel B: tests of H3b
Learned Trust Dispositional Trust Satisfaction_Index 0.267 [0.108,0.442]* p< 0.05
Learned Trust Dispositional Trust Serv_Efficiency 0.188 [0.029, 0.390]* p< 0.05
Learned Trust Dispositional Trust Serv_Responsiveness 0.110 [-0.074, 0.274] p> 0.05
Learned Trust Dispositional Trust Serv _Empathy 0.264 [0.074, 0.429]* p< 0.05
Learned Trust Dispositional Trust Serv_Reliability 0.117 [0.002, 0.282]* p< 0.05
Learned Trust Dispositional Trust Serv_Assurances 0.299 [0.091, 0.481]* p< 0.05
Learned Trust Dispositional Trust Provider Loyalty 0.218 [0.021, 0.407]* p< 0.05
Learned Trust Dispositional Trust Recomm_Acceptance 0.272 [0.080, 0.477]* p< 0.05

Notes: *p < 0.05.

9
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

Table 4
Regressions of customer outcome variables on the (high or low) level of situational trust in the chatbot service-provider. These analyses are based on Study 2.
IVs/DVs = − 1 − 2 − 3 − 4 − 5 − 6 − 7 − 8

Sat_Index Serv_Efficiency Serv_Respons. Serv _Empathy Serv_Reliability Serv_Assurances Provider Loyalty Recomm_Accept

Panel A: Situational Trust


High-Situational- Trust 0.856*** 0.607*** 0.457** 0.610** 0.648*** 0.802*** 0.918*** 0.801***
(0.166) (0.165) (0.148) (0.188) (0.160) (0.183) (0.203) (0.223)
Experience w/Chatbotsa 0.0994 0.101 0.0937 0.171* 0.0468 0.11 0.166* − 0.0396
(0.068) (0.067) (0.060) (0.077) (0.065) (0.074) (0.082) (0.091)
Experience w/Human CSRa 0.11 0.0463 0.0367 0.0338 0.0336 0.0868 0.0641 − 0.206*
(0.068) (0.067) (0.061) (0.077) (0.065) (0.075) (0.083) (0.091)
Involvement − 0.0303 − 0.109 0.0545 − 0.116 − 0.0701 − 0.135 − 0.0837 0.304*
(0.109) (0.108) (0.098) (0.124) (0.105) (0.120) (0.133) (0.147)
Disp. Trust in People 0.154 0.152 0.136 0.182 0.197* 0.207* 0.203* 0.154
(0.082) (0.081) (0.073) (0.093) (0.079) (0.091) (0.100) (0.110)
Constant 2.708*** 3.982*** 3.759*** 3.201*** 3.229*** 3.173*** 2.298** 1.797
(0.695) (0.689) (0.621) (0.789) (0.668) (0.766) (0.848) (0.933)
adj. R2 0.142 0.081 0.062 0.081 0.086 0.117 0.121 0.088
F 8.204 4.82 3.905 4.848 5.106 6.763 6.993 5.204
p 0.000 0.000 0.002 0.000 0.000 0.000 0.000 0.000
N 219 219 219 219 219 219 219 219
Panel B: Extended model
High-Situational- Trust 0.798*** 0.573*** 0.425** 0.572*** 0.605*** 0.771*** 0.845*** 0.753***
(0.153) (0.154) (0.132) (0.170) (0.138) (0.160) (0.182) (0.224)
Disp. Trust in Service Tech 0.194* 0.137 0.137 0.207* 0.252*** 0.280** 0.284** 0.0297
(0.082) (0.082) (0.070) (0.091) (0.074) (0.085) (0.097) (0.119)
Learned Trust in Chatbots 0.291*** 0.296*** 0.343*** 0.383*** 0.327*** 0.340*** 0.368*** 0.0427
(0.075) (0.076) (0.065) (0.083) (0.068) (0.078) (0.090) (0.110)
Experience w/Chatbotsa 0.0291 0.042 0.0504 0.113 − 0.0353 0.00978 0.0461 − 0.00837
(0.066) (0.067) (0.057) (0.073) (0.060) (0.069) (0.079) (0.097)
Experience w/Human CSRa 0.0912 0.0445 0.0409 0.0268 0.0181 0.0627 0.043 − 0.241**
(0.063) (0.063) (0.054) (0.070) (0.057) (0.066) (0.075) (0.092)
Involvement 0.0917 0.0164 0.184* 0.0318 0.0636 0.00487 0.0526 0.332*
(0.104) (0.104) (0.089) (0.115) (0.093) (0.108) (0.124) (0.152)
Disp. Trust in People 0.0412 0.0623 0.0271 0.0464 0.0713 0.0848 0.0634 0.0876
(0.077) (0.078) (0.067) (0.086) (0.070) (0.081) (0.092) (0.113)
Constant 1.054 2.278** 1.817** 0.954 1.212 1.057 0.252 1.688
(0.703) (0.708) (0.606) (0.779) (0.633) (0.734) (0.837) (1.027)
adj. R2 0.281 0.214 0.284 0.277 0.332 0.328 0.295 0.083
F 12.85 9.234 13.01 12.62 16.05 15.77 13.68 3.754
p 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.001
N 213 213 213 213 213 213 213 213

5.3. Study 3 the value of one if the reservation was handled by a human agent or
zero—by a chatbot-human service team.
Finally, Study 3 evaluated two approaches to stimulating trust in Looking first at the results based on the sample that was primed for
chatbots in order to achieve customer outcome parity with human ser­ low trust in Stella, we note that all the regressions are statistically sig­
vice. The first approach entailed priming situational trust in the chatbot nificant. However, the coefficient on the HM_Service dummy is signifi­
providing service. To test our hypothesis, we regressed our dependent cantly positive in every regression (at p < 0.05 and lower). Thus, all the
variables on the dummy variable labeled, HM_Service, which took the evidence strongly supports the “inferiority” hypothesis H6a. Stated
value of one for the human-agent service condition and zero for chatbot differently, we find no evidence that a brief human intervention
service. We show the results in Panel A of Table 5. While the coefficient following chatbot service can meaningfully improve customer outcomes
on the focal dummy was not statistically significant in four regressions, when consumers have low situational trust in the chatbot service-
it was significantly positive (at p < 0.05) in four other regressions provider.
involving Satisfaction_Index, Service_Empathy, Service_Assurances, and Next considering the sample with high situational trust in the chatbot
Repatronage. We note that this pattern of results makes intuitive sense, in Stella, we note that the coefficient on the HM_Service dummy is signifi­
particular, the lack of impact on perceived empathy. On balance, our cantly positive only in the regression of Service_Empathy, meaning that
data provide partial support for H5, which is framed as a non-inferiority human service outperformed the chatbot-human agent combination
test. It posits that even when primed for high situational trust in the only on that customer outcome. The dummy coefficient in the other
chatbot service giver, individuals report worse outcomes compared with seven regressions is not statistically significant at p < 0.05. Although, we
equivalent service by a human agent. We find this to be true with respect note that the focal coefficient is marginally significant at p < 0.1 and
to half of our dependent variables. Reframing our results in a positive negative in the regression of Service_Efficiency, suggesting that the
light, one can, alternatively, state that priming particularized situational chatbot-human agent combination borderline outperformed the human
trust in the chatbot service-giver helped to achieve non-inferiority with service condition in terms of perceived service efficiency. These two
human service on half of the outcome measures. results align with the basic intuition that human agents will tend to score
We next evaluated our second proposed approach to stimulating higher on empathy in most circumstances, whereas chatbots are
trust in chatbot service. This approach involved comparing human ser­ generally associated with service efficiency. In sum, our data lead us to
vice with a chatbot-human combination primed for low or high situa­ reject the inferiority hypotheses H6b for seven out of our eight depen­
tional trust. We estimated our regressions in two samples which dent variables. Stated differently, these results largely support our
contained 212 and 203 valid observations, respectively. In both samples, proposition that a chatbot-human-agent combination, as described in
we regressed our battery of dependent variables on a dummy which took this study, can match human-agent service on customer outcomes

10
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

Table 5
Regressions of the customer outcome variables on the service provider type—human agent vs. chatbot/human agent—in the subsets with low or high situational trust
in the chatbot. These analyses are based on Study 3.
IVs/DVs = − 1 − 2 − 3 − 4 − 5 − 6 − 7 − 8

Sat_Index Serv_Efficiency Serv_Respons. Serv _Empathy Serv_Reliability Serv_Assurances Provider Loyalty Recomm_Accept

Panel A: HM vs. Chtabot w/High Trust


HM_Servicea 0.351* 0.0284 0.246 0.447* 0.153 0.373* 0.473* 0.218
(0.16) (0.18) (0.16) (0.18) (0.18) (0.18) (0.21) (0.27)
b
Experience w/Chatbots 0.0264 0.0603 0.0256 0.125 0.032 0.0695 0.0951 − 0.108
(0.06) (0.07) (0.06) (0.07) (0.07) (0.07) (0.08) (0.10)
b
Experience w/Human CSR − 0.00782 − 0.0559 − 0.0288 0.00376 − 0.0203 0.00996 − 0.0224 − 0.240*
(0.06) (0.07) (0.06) (0.07) (0.07) (0.07) (0.08) (0.10)
Involvement 0.0281 − 0.0303 0.0542 0.0596 0.0557 0.102 0.0128 − 0.0122
(0.07) (0.08) (0.07) (0.08) (0.08) (0.08) (0.10) (0.12)
Disp. Trust in People 0.169* 0.265** 0.201** 0.242** 0.265** 0.207* 0.208* 0.119
(0.08) (0.09) (0.07) (0.09) (0.08) (0.08) (0.10) (0.13)
adj. R2 0.037 0.028 0.042 0.079 0.042 0.066 0.036 0.026
F 2.541 2.14 2.77 4.451 2.766 3.853 2.498 2.062
p 0.0296 0.0624 0.0192 0.000731 0.0194 0.00237 0.0321 0.0718
N 202 202 202 202 202 202 202 202
Panel B: HM vs. Chatbot-HM w/Low Trust
HM_Service 0.618*** 0.376* 0.466** 0.965*** 0.566*** 0.969*** 0.776*** 0.522*
(0.15) (0.16) (0.14) (0.17) (0.17) (0.18) (0.20) (0.25)
Experience w/Chatbots 0.0389 − 0.034 − 0.0141 − 0.0561 − 0.0642 − 0.0704 0.0993 − 0.161
(0.06) (0.06) (0.06) (0.07) (0.07) (0.07) (0.08) (0.10)
Experience w/Human CSR − 0.0436 − 0.028 − 0.0148 0.0444 − 0.0511 − 0.0458 − 0.0728 − 0.194*
(0.06) (0.06) (0.05) (0.07) (0.06) (0.07) (0.08) (0.09)
Involvement 0.0547 0.0473 0.0731 0.151* 0.0584 0.115 0.164* − 0.0205
(0.06) (0.06) (0.05) (0.06) (0.06) (0.07) (0.08) (0.09)
Disp. Trust in People 0.331*** 0.353*** 0.281*** 0.295*** 0.384*** 0.391*** 0.347*** 0.0379
(0.07) (0.08) (0.07) (0.09) (0.08) (0.09) (0.10) (0.12)
adj. R2 0.163 0.105 0.125 0.204 0.145 0.219 0.156 0.043
F 9.195 5.94 7.022 11.78 8.128 12.85 8.813 2.889
p 6.48E-08 0.0000373 0.00000441 4.95E-10 0.000000508 6.97E-11 0.000000135 0.0152
N 212 212 212 212 212 212 212 212
Panel C: HM vs. Chatbot-HM w/High Trust
HM_Service − 0.0487 − 0.266 0.0968 0.443* − 0.0859 0.195 − 0.00616 0.232
(0.15) (0.16) (0.14) (0.17) (0.16) (0.16) (0.20) (0.26)
Experience w/Chatbots 0.0167 0.00363 − 0.00819 0.0353 − 0.0193 − 0.0318 0.0484 − 0.198
(0.06) (0.07) (0.06) (0.07) (0.07) (0.07) (0.08) (0.11)
Experience w/Human CSR 0.1 0.0916 0.0329 0.13 0.0444 0.0601 0.109 − 0.0391
(0.06) (0.06) (0.05) (0.07) (0.06) (0.06) (0.08) (0.10)
Involvement 0.101 0.116* 0.174*** 0.257*** 0.217*** 0.228*** 0.204** − 0.101
(0.05) (0.06) (0.05) (0.06) (0.06) (0.06) (0.07) (0.09)
Disp. Trust in People − 0.00372 0.0521 0.0674 0.105 0.108 0.0979 0.00896 − 0.00583
(0.07) (0.07) (0.06) (0.08) (0.07) (0.07) (0.09) (0.12)
adj. R2 0.024 0.035 0.069 0.168 0.08 0.103 0.052 0.015
F 1.996 2.458 3.983 9.134 4.501 5.618 3.207 1.611
p 0.0809 0.0347 0.00183 7.86E-08 0.00066 0.0000727 0.00832 0.159
N 203 203 203 203 203 203 203 203

Notes: All the regressions included a constant which is not shown.


Standard errors are in parentheses.
*p < 0.05, **p < 0.01, ***p < 0.001.
a
HM_Service designates the human-agent service condition.
b
Experience with chatbots/human customer service representatives (CSR) in the previous two years.

provided the customers have situational trust in the chatbot. from satisfied customers. Moreover, a reputation for superior customer
service can bestow a competitive advantage, as exemplified by com­
6. Discussion, managerial implications, and conclusions panies like Nordstrom’s and Apple (Garg, 2020). Hence,
customer-oriented organizations will likely seek ways to implement
Technology analysts proclaim that chatbots are rapidly becoming “a chatbots, as well as supporting structures and activities, in a manner that
critical technology component of a service organization’s self-service can help close the satisfaction gap between human and automated
strategy” (Gartner, Inc. 2023). At the same time, industry surveys customer service.
reveal that consumers generally prefer service by human representa­ We draw on our trust framework to propose two chatbot service
tives. We confirm this preference in our research setting. Compared with configurations that may stimulate customer trust during a service
equivalent service provided by a chatbot, our research participants re­ encounter and, thereby, help improve consumer experience with chat­
ported higher satisfaction with human customer service. They also bot service. We empirically evaluate both approaches on their ability to
demonstrated greater provider repatronage intent and recommendation match human service on customer satisfaction, service agent’s recom­
acceptance after a service encounter with a human agent. mendation acceptance, and repatronage intentions. The first approach
The tension between service providers’ chatbot adoption and con­ entails priming situational trust in the chatbot service provider at the
sumer hesitancy in accepting chatbot service presents a challenge for the point-of-service. We find this approach to be more effective than basic
adopting organizations in the long run. Many companies and non-profits chatbot service (without priming), although, falling short of human
strive for positive customer outcomes and marketing benefits which flow service on half of the customer outcome measures used in our tests. Our

11
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

second proposed approach combines situational trust priming, as in the 2019).


first approach, with a limited human intervention at the conclusion of a Finally, companies should consider sponsoring industry-wide ini­
service encounter which is handled by a chatbot. Adding a human tiatives that can directly stimulate chatbot acceptance. In particular,
element is intended to further stimulate situational trust in a chatbot via developing learned trust in chatbot service provision can have a positive
the halo effect, and also additively increase overall trust in the service impact on important customer outcomes. While experience-based
“team,” since consumers have higher dispositional and learned trust in learned trust will grow organically over time, as consumers are
human service providers. Our empirical analysis shows that the second increasingly served by evolving virtual agents, the process may be
approach is not only more effective than the first one, but it can also further accelerated by stimulating other common learning routes, such
match human service on the focal customer outcomes. as using trade association communications and public relations to
educate consumers on the inner workings, safety, integrity, benevo­
6.1. Managerial implications lence, and successes of service chatbots.

Our experimental results corroborate industry surveys which point to 6.2. Theoretical implications
low consumer trust in chatbots as an impediment to chatbots’ wide­
spread consumer acceptance and continued use. The development of AI This current research builds on previous research on trust in service
chatbots’ problem-solving and relational capabilities is likely to provide technologies (i.e. Mayer et al., 1995; Glikson and Woolley, 2020). It is
an effective solution to the trust conundrum in the long run. In the unique in that it combines theoretical lenses from marketing (i.e. Auh,
interim, we contend that marketing interventions, such as trust priming, 2005), information systems (i.e. McKnight et al., 2011) and human
and efficient chatbot-human service combinations may provide an factors research (i.e. Hoff and Bashir, 2015), which allows us to develop
improvement on service configurations that solely feature live or chat­ new approaches and insights. We propose a novel theoretical framework
bot agents. focusing on the role of trust in AI service encounters. We use the
To begin with, consumers vary in their preference for chatbot or framework to generate empirical evidence showing different outcomes
human service, as well as their dispositional and learned trust in chat­ involving consumer response to chatbot service based on the interplay of
bots. Hence, segmentation- and targeting-based approaches may present three distinct layers of trust: dispositional, learned and situational (i.e.
a low-hanging fruit on the path to improving chatbot-driven customer Marsh and Dibben, 2003). Hence, our research advances the trust
outcomes. For example, companies may leverage self-segmentation by literature by proposing a model of trust that is specifically adapted to
offering consumers a choice of chatbot or human service provider at the consumer interactions with chatbots. Another important theoretical
point-of-service (possibly, with the option of changing at any time be­ distinction of our framework is in presenting a dynamic view of trust
tween the two, to facilitate chatbot acceptance). Further self-targeting which accounts for changes in consumer trust over time. This opens new
may be achieved by enabling different, even a-la-carte, configurations avenues for research into temporal aspects of trust, in contrast to the
within chatbot service offerings (e.g., different chatbot personas). Firms prevalent approaches to trust that focus on trustee attributes, such as
that use chatbot service can also consider adding the trust variables as confidence, integrity, reliability, and benevolence.
additional inputs into their segmentation analysis. Further, our research provides insights into the process of trust for­
Another promising chatbot implementation approach, which may be mation and elucidates the relationships between the focal trust layers. At
combined with self-segmented offerings, includes taking steps to stim­ the most fundamental level, one’s interactions with chatbots are shaped
ulate particularized situational trust in the chatbot service-giver. This by the extent of one’s propensity to trust service technologies—an
would include careful orchestration of contextual cues that can enhance assumption that these technologies are largely benign, beneficial, and
consumer confidence in the chatbot’s abilities (e.g., based on its success reliable. Over time, positive experiences with customer service chatbots
rate), honesty and benevolence. Conceptually and ethically, this would and other learning behaviors stimulate the development of generalized
not be different from the existing practice of human service providers learned trust in chatbots (i.e. Kong et al., 2020). In addition, during
displaying in their offices and on their marketing materials their degree service encounters with a chatbot, consumers may be exposed to
diplomas, testimonials, and prominent endorsements. We note that, contextual cues highlighting the chatbot’s capabilities, track record, or
while this approach was not completely on a par with human service in performance guarantees (i.e. Chinchanachokchai et al., 2021). These
our tests, it is very low-costand can be further enhanced using sophis­ cues motivate particularized situational trust in the specific chatbot
ticated priming instruments. service-giver.
However, the most promising and best-performing chatbot configu­ Our empirical investigation confirms that these trust layers impact
ration in our tests involved combining a chatbot in the primary service consumer experience with chatbots. To our knowledge we are the first to
role with a human agent in a supporting role. This approach naturally provide empirical evidence that all three trust layers have a direct effect
instills higher trust in the service team, although it performs better when on customer satisfaction with chatbot service and subsequent trusting
combined with a prime to enhance situational trust in the chatbot. To behaviors, including recommendation acceptance and repatronage in­
the extent that a human agent can step in following the conclusion of a tentions. Dispositional trust in service technologies, in particular, was
service encounter, for example, for quality control and reassurance, the found to play a central role by mediating the effect of learned trust in
operating cost of such configuration may be only marginally higher than chatbots. This result conforms with the theoretical prediction that “ob­
that of fully automated service. To reduce costs further, it can be servations are theory-laden” (Govier, 1994, p. 244) and trust propensity
selectively applied to non-routine service encounters. continues to impact one’s trusting behavior even in the presence of
We believe that the outlined chatbot-human implementation with a established trust.
situational trust prime represents a reasonable path towards both more
efficient and effective customer service that has many of the advantages 6.3. Limitations and future research
of chatbot and human agents, while requiring acceptable trade-offs.
Although it may be somewhat more expensive than full automation, Our research includes some limitations. First, while we designed our
and poses scaling challenges to incorporate the human element, it studies in such a way to limit potential confounds, we acknowledge that
promises superior customer satisfaction and important marketing ben­ there are additional variables which may influence the perception of the
efits, including incremental sales. If nothing else, the proposed config­ chatbot. For example, although we collected some demographic vari­
uration may serve as a functional bridge towards full automation in the ables in our studies (i.e., gender and age), and verified that our results
future when consumers can be served by virtual agents endowed with are robust to their inclusion into the regression models, we did not have
high general and emotional intelligence (Prentice, 2019; Prentice et al., information on other additional socio-demographic variables such as

12
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

income, education and occupation. Culture is another dimension for Basso, K., Pizzutti, C., 2016. Trust recovery following a double deviation. J. Serv. Res. 19
(2), 209–223.
which we did not have control. Given that culture is not a typical control
Beaner, L., 2022. Two Thirds of Americans Feel Secure about Their Purchases when
variable in the literature, we leave it to further investigation on specific Dealing with a Live Person. SWNS. URL: https://swnsdigital.com/us/2022/10/t
service contexts. wo-thirds-of-americans-feel-secure-about-their-purchases-when-dealing-with-a-live-
Additionally, the demographics (race, age and gender) of the tour person/. Accessed on 5/9/2023.
Bowman, S.R., 2023. Eight Things to Know about Large Language Models. arXiv preprint
guides and service providers may have influenced the participants. Since arXiv:2304.00612.
the tour guide pictures and the chatbot avatar were the same across Benamati, J., Fuller, M.A., Serva, M.A., Baroudi, J., 2009. Clarifying the integration of
conditions (Web Appendix B), it is unlikely that any effects would in­ trust and TAM in e-commerce environments: implications for systems design and
management. IEEE Trans. Eng. Manag. 57 (3), 380–393.
fluence the results. However, future research should consider comparing Biros, D.P., Fields, G., Gunsch, G., 2003. The effect of external safeguards on human-
different demographics between conditions, and consider the possibility information system trust in an information warfare environment. In: Proceedings of
of gender influence. the 36th Annual Hawaii International Conference on System Sciences. IEEE,
pp. 1–10.
Next, consumers sometimes have a choice between using a chatbot Burke, R.R., 2002. Technology and the customer interface: what consumers want in the
immediately or queueing up for a human agent. This was not an option physical and virtual store. J. Acad. Market. Sci. 30 (4), 411–432.
in our study. We believe our setting is commonplace in this regard. At Chen, S.C., Dhillon, G.S., 2003. Interpreting dimensions of consumer trust in e-
commerce. Inf. Technol. Manag. 4 (2/3), 303–318.
the same time, we recognize that enabling the choice of a service-giver Chen, Q., Gong, Y., Lu, Y., Tang, J., 2022. Classifying and measuring the service quality
may affect one’s subsequent experiences. Since that various chatbot of AI chatbot in frontline service. J. Bus. Res. 145, 552–568.
service options are possible, as discussed in the preceding section, it Chinchanachokchai, S., Thontirawong, P., Chinchanachokchai, P., 2021. A tale of two
recommender systems: the moderating role of consumer expertise on artificial
would be instructive for future research to examine consumer trust
intelligence based product recommendations. J. Retailing Consum. Serv. 61,
patterns and related experiences in light of the distinct paths that con­ 102528.
sumers may take leading up to their service encounter. Chu, P., Lee, G., Chao, Y., 2012. Service quality, customer satisfaction, customer trust,
In the same vein, we contextualize this research in diverse customer and loyalty in an e-banking context. SBP (Soc. Behav. Pers.) 40 (8), 1271–1284.
Chung, M., Ko, E., Joung, H., Kim, S.J., 2020. Chatbot e-service and customer satisfaction
service applications that entail service provided by human and chatbot regarding luxury brands. J. Bus. Res. 117, 587–595.
agents via online chat. The degree to which our recommendations may Ciechanowski, L., Przegalinska, A., Magnuski, M., Gloor, P., 2019. In the shades of the
extend to face-to-face service contexts (e.g., involving humanoid ro­ uncanny valley: an experimental study of human–chatbot interaction. Future
Generat. Comput. Syst. 92, 539–548.
bots), or voice-based interactions, which, both, enable a distinct set of Clark, S., 2022. Do your customers trust your AI? Downloaded from. https://www.cmswi
cues and limitations would need to be established in future research. re.com/customer-experience/do-your-customers-trust-your-ai/. Accessed on 01/10/
Another future research direction would be to explore the boundary 2023.
Cronin Jr., J.J., Taylor, S.A., 1994. SERVPERF versus SERVQUAL: reconciling
conditions for chatbot service acceptance. For example, we pursue our performance-based and perceptions-minus-expectations measurement of service
research questions in a rather general setting that embeds moderate quality. J. Market. 58 (1), 125–131.
level of risk, and focuses on a mostly positive consumer experi­ Doney, P.M., Cannon, J.P., 1997. An examination of the nature of trust in buyer–seller
relationships. J. Market. 61 (1), 35–51.
ence—making a travel reservation. However, consumers often contact Dwyer, F.R., Schurr, P.H., Oh, S., 1987. Developing buyer-seller relationships. J. Market.
customer service with a problem (e.g., a missed flight, a product mal­ 51 (2), 11–27.
function, a return, a missing package, etc.). While our framework and Elliott, K.M., 1994. SERVPERF versus SERVQUAL: a marketing management dilemma
when assessing service quality. J. Market. Manag. 4 (2), 56–61.
recommendations generalize to chatbot service experiences in such
Evans, A.M., Revelle, W., 2008. Survey and behavioral measurements of interpersonal
settings, further empirical research in distinct settings is warranted. trust. J. Res. Pers. 42 (6), 1585–1593.
Ferris, G.R., Judge, T.A., Rowland, K.M., Fitzgibbons, D.E., 1994. Subordinate influence
and the performance evaluation process: test of a model. Organ. Behav. Hum. Decis.
Declaration of competing interest
Process. 58 (1), 101–135.
Freitag, M., Traunmüller, R., 2009. Spheres of trust: an empirical analysis of the
The authors declare that they have no known competing financial foundations of particularised and generalised trust. Eur. J. Polit. Res. 48 (6),
782–803.
interests or personal relationships that could have appeared to influence
Fukuyama, F., 1995. Dust: the Social Virtues and the Creation of Prosperity. Hamish
the work reported in this paper. Hamilton.
Gabrieli, G., Ng, S., Esposito, G., 2021. Hacking trust: the presence of faces on automated
Data availability teller machines (ATMs) affects trustworthiness. Behav. Sci. 11 (6), 91.
Garbarino, E., Johnson, M.S., 1999. The different roles of satisfaction, trust, and
commitment in customer relationships. J. Market. 63 (2), 70–87.
Data will be made available on request. Garg, P., 2020. How has Apple gained competitive edge over the years? Downloaded
from. https://medium.com/macoclock/how-has-apple-gained-competitive-edge-ove
r-the-years-aff0a259de0d. Accessed on 01/02/2023.
Appendix A. Supplementary data Gartner, Inc., 2018. Gartner Survey Finds that Most Consumers Have Underwhelming
Digital Experience. Available from: https://www.gartner.com/en/newsroo
Supplementary data to this article can be found online at https://doi. m/press-releases/2018-10-03-gartner-survey-finds-that-most-consumers-have-under
whelming-digital-experiences. Accessed on 10/14/2022.
org/10.1016/j.jretconser.2023.103600. Gartner, 2023. The Customer Service Chatbot Deployment Guide. Inc.. https://www.gart
ner.com/en/customer-service-support/trends/customer-service-chatbot-guide.
References Accessed on 01/01/2023.
Gefen, D., 2000. E-commerce: the role of familiarity and trust. Omega 28 (5), 725–737.
Geyskens, I., Steenkamp, J.B.E., Kumar, N., 1998. Generalizations about trust in
Alba, J.W., Hutchinson, J.W., 1987. Dimensions of consumer expertise. J. Consum. Res.
marketing channel relationships using meta-analysis. Int. J. Res. Market. 15 (3),
13 (4), 411–454.
223–248.
Aleedy, M., Shaiba, H., Bezbradica, M., 2019. Generating and analyzing chatbot
Glanville, J.L., Shi, Q., 2020. The Extension of particularized trust to generalized and
responses using natural language processing. Int. J. Adv. Comput. Sci. Appl. 10 (9),
out-group trust: the constraining role of collectivism. Soc. Forces 98 (4), 1801–1828.
60–68.
Glikson, E., Woolley, A.W., 2020. Human trust in artificial intelligence: review of
Ameen, N., Tarhini, A., Reppel, A., Anand, A., 2021. Customer experiences in the age of
empirical research. Acad. Manag. Ann. 14 (2), 627–660.
artificial intelligence. Comput. Hum. Behav. 114, 106548.
Govier, T., 1994. Is it a jungle out there? Trust, distrust and the construction of social
Amin-Naseri, M., Gilbert, S., 2014. A system dynamics approach to building team trust
reality. Dialogue 33, 237–252.
models: exploring the challenges. In: Pedagogy that Makes A Difference: Exploring
Gregori, N., Daniele, R., Altinay, L., 2014. Affiliate marketing in tourism: determinants of
Domain-independent Principles across. Instructional Management Research within
consumer trust. J. Trav. Res. 53 (2), 196–210.
the ITS Community, p. 49.
Harborth, D., Pape, S., 2021. Investigating privacy concerns related to mobile augmented
Auh, S., 2005. The effects of soft and hard service attributes on loyalty: the mediating
reality Apps–A vignette based online experiment. Comput. Hum. Behav. 122,
role of trust. J. Serv. Market. 19 (2), 81–92.
106833.
Barber, B., 1983. The Logic and Limits of Trust. Rutgers University Press.
Hasan, R., Shams, R., Rahman, M., 2021. Consumer trust and perceived risk for voice-
Bartneck, C., Kulić, D., Croft, E., Zoghbi, S., 2009. Measurement instruments for the
controlled artificial intelligence: the case of Siri. J. Bus. Res. 131, 591–597.
anthropomorphism, animacy, likeability, perceived intelligence, and perceived
safety of robots. International Journal of Social Robotics 1, 71–81.

13
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600

Hayes, A.F., 2018. PROCESS: A Versatile Computational Tool for Observed Variable Parasuraman, A., Zeithaml, V.A., Berry, L.L., 1985. A conceptual model of service quality
Mediation, Moderation, and Conditional Process Modeling. and its implications. J. Market. 49 (4), 41–50.
Hill, J., Ford, W.R., Farreras, I.G., 2015. Real conversations with artificial intelligence: a Pavlou, P.A., 2003. Consumer acceptance of electronic commerce: integrating trust and
comparison between human–human online conversations and human–chatbot risk with the technology acceptance model. Int. J. Electron. Commer. 7 (3), 101–134.
conversations. Comput. Hum. Behav. 49, 245–250. Pegasystems, 2019. What Consumers Really Think about AI: A Global Study. Retrieved
Ho, G., Wheatley, D., Scialfa, C.T., 2005. Age differences in trust and reliance of a from. https://www.ciosummits.com/what-consumers-really-think-about-ai.pdf.
medication management system. Interact. Comput. 17, 690–710. Pegoraro, A., Kumari, K., Fereidooni, H., Sadeghi, A.R., 2023. To ChatGPT, or Not to
Hoff, K.A., Bashir, M., 2015. Trust in automation: integrating empirical evidence on ChatGPT: that Is the Question!. arXiv preprint arXiv:2304.01487.
factors that influence trust. Hum. Factors 57 (3), 407–434. Perez, E., Ringer, S., Lukošiūtė, K., Nguyen, K., Chen, E., Heiner, S., et al., 2022.
Hsu, C.L., Lin, J.C.C., 2023. Understanding the user satisfaction and loyalty of customer Discovering Language Model Behaviors with Model-Written Evaluations. arXiv
service chatbots. J. Retailing Consum. Serv. 71, 103211. preprint arXiv:2212.09251.
Internova Travel Group, 2021. The Future of Travel Is Human. Downloaded from. htt Pizzi, G., Scarpi, D., Pantano, E., 2021. Artificial intelligence and the new forms of
ps://internova.com/the-future-of-travel-is-human/. Accessed on 5/3/2023. interaction: who has the control when interacting with a chatbot? J. Bus. Res. 129,
Jabeur, S.B., Ballouk, H., Arfi, W.B., Sahut, J.M., 2023. Artificial intelligence applications 878–890.
in fake review detection: bibliometric analysis and future avenues for research. Prentice, C., 2019. Managing service encounters with emotional intelligence. J. Retailing
J. Bus. Res. 158, 113631. Consum. Serv. 51, 344–351.
Jacoby, J., Chesnut, R.W., 1978. Brand Loyalty: Measurement and Management. John Prentice, C., Wang, X., Loureiro, S.M.C., 2019. The influence of brand experience and
Wiley & Sons, New York, NY. service quality on customer engagement. J. Retailing Consum. Serv. 50, 50–59.
Jörling, M., Böhm, R., Paluch, S., 2019. Service robots: Drivers of perceived Rau, P.L., Li, Y., Li, D., 2009. Effects of communication style and culture on ability to
responsibility for service outcomes. J. Service Res. 22 (4), 404–420. accept recommendations from robots. Comput. Hum. Behav. 25, 587–595.
Kennedy, M.S., Ferrell, L.K., LeClair, D.T., 2001. Consumers’ trust of salesperson and Rese, A., Ganster, L., Baier, D., 2020. Chatbots in retailers’ customer communication:
manufacturer: an empirical study. J. Bus. Res. 51 (1), 73–86. how to measure their acceptance? J. Retailing Consum. Serv. 56, 102176.
Kong, Y., Wang, Y., Hajli, S., Featherman, M., 2020. In sharing economy we trust: Roose, K., 2023. A Conversation with Bing’s Chatbot Left Me Deeply Unsettled. New York
examining the effect of social and technical enablers on millennials’ trust in sharing Times. Retrieved from. https://www.nytimes.com/2023/02/16/technology/bing-c
commerce. Comput. Hum. Behav. 108, 105993. hatbot-microsoft-chatgpt.html.
Lee, J.D., See, K.A., 2004. Trust in automation: designing for appropriate reliance. Hum. Rodrigues, L.L., Barkur, G., Varambally, K.V.M., Motlagh, F.G., 2011. Comparison of
Factors 46 (1), 50–80. SERVQUAL and SERVPERF metrics: an empirical study. The TQM Journal 23 (6),
Lewis, J.D., Weigert, A., 1985. Trust as a social reality. Soc. Forces 63, 967–985. 629–643.
Li, S., Peluso, A.M., Duan, J., 2023. Why do we prefer humans to artificial intelligence in Sahoo, N., Krishnan, R., Duncan, G., Callan, J., 2012. Research note—the Halo effect in
telemarketing? A mind perception explanation. J. Retailing Consum. Serv. 70, multicomponent ratings and its implications for recommender systems: the case of
103139. yahoo! Movies. Inf. Syst. Res. 23 (1), 231–246.
Lin, Y.T., Doong, H.S., Eisingerich, A.B., 2021. Avatar design of virtual salespeople: Sands, S., Ferraro, C., Campbell, C., Tsao, H.Y., 2021. Managing the human–chatbot
mitigation of recommendation conflicts. J. Serv. Res. 24 (1), 141–159. divide: how service scripts influence service experience. J. Serv. Manag. 32 (2),
Madhavan, P., Phillips, R.R., 2010. Effects of computer self-efficacy and system 246–264.
reliability on user interaction with decision support systems. Comput. Hum. Behav. Serva, M.A., Benamati, J.S., Fuller, M.A., 2005. Trustworthiness in B2C e-commerce: an
26, 199–204. examination of alternative models. ACM SIGMIS - Data Base: the DATABASE for
Mariani, M.M., Hashemi, N., Wirtz, J., 2023. Artificial intelligence empowered Advances in Information Systems 36 (3), 89–108.
conversational agents: a systematic literature review and research agenda. J. Bus. Sheehan, B., Jin, H.S., Gottlieb, U., 2020. Customer service chatbots: anthropomorphism
Res. 161, 113838. and adoption. J. Bus. Res. 115, 14–24.
Marsh, S., Dibben, M.R., 2003. The role of trust in information science and technology. Shumanov, M., Johnson, L., 2021. Making conversations with chatbots more
Annu. Rev. Inform. Sci. Technol. 37, 465–498. personalized. Comput. Hum. Behav. 117, 106627.
Mayer, R.C., Davis, J.H., Schoorman, F.D., 1995. An integrative model of organizational Söderlund, M., 2006. Measuring customer loyalty with multi-item scales: a case for
trust. Acad. Manag. Rev. 20 (3), 709–734. caution. Int. J. Serv. Ind. Manag. 17 (1), 76–98.
McKnight, D.H., Choudhury, V., Kacmar, C., 2002. Developing and validating trust Ullman, D., Malle, B.F., 2017. Human-robot trust: just a button press away. In:
measures for e-commerce: an integrative typology. Inform. Syst. Res. 13 (3), Proceedings of the Companion of the 2017 ACM/IEEE International Conference on
334–359. Human-Robot Interaction - HRI ’17, pp. 309–310.
McKnight, D.H., Cummings, L.L., Chervany, N.L., 1998. Initial trust formation in new van Wezel, M., Croes, E.A., Antheunis, M.L., 2021. "I’m here for you": can social chatbots
organizational relationships. Acad. Manag. Rev. 23 (3), 473–490. truly support their users? A literature review. In: International Workshop on Chatbot
McKnight, D.H., Carter, M., Thatcher, J.B., Clay, P.F., 2011. Trust in a specific Research and Design. Springer, Cham, pp. 96–113.
technology: an investigation of its components and measures. ACM Transactions on Wang, X., Lin, X., Shao, B., 2022. How does artificial intelligence create business agility?
Management Information Systems (TMIS) 2 (2), 1–25. Evidence from chatbots. Int. J. Inf. Manag. 66, 102535.
McLean, G., Osei-Frimpong, K., Barhorst, J., 2021. Alexa, do voice assistants influence Wiggers, K., 2023. IBM Intros a Slew of New AI Services, Including Generative Models.
consumer brand engagement?–Examining the role of AI powered voice assistants in TechCrunch. Retrieved from. https://techcrunch.com/2023/05/09/ibm-intros-a-sle
influencing consumer brand engagement. J. Bus. Res. 124, 312–328. w-of-new-ai-services-including-generative-models/.
Meuter, M.L., McCabe, D.B., Curran, J.M., 2013. Electronic word-of-mouth versus Xu, K., Chen, X., Huang, L., 2022. Deep mind in social responses to technologies: a new
interpersonal word-of-mouth: are all forms of word-of-mouth equally influential? approach to explaining the Computers are Social Actors phenomena. Comput. Hum.
Serv. Market. Q. 34 (3), 240–256. Behav. 134, 107321.
Morgan, R.M., Hunt, S.D., 1994. The commitment-trust theory of relationship marketing. Zaichkowsky, J.L., 1985. Measuring the involvement construct. J. Consum. Res. 12 (3),
J. Market. 58 (3), 20–38. 341–352.
Nass, C., Moon, Y., 2000. Machines and mindlessness: social responses to computers. Zhang, J., 2010. The sound of silence: observational learning in the US kidney market.
J. Soc. Issues 56 (1), 81–103. Market. Sci. 29 (2), 315–335.
Nisbett, R.E., Wilson, T.D., 1977. The halo effect: evidence for unconscious alteration of
judgments. J. Pers. Soc. Psychol. 35 (4), 250–256.

14

You might also like