Professional Documents
Culture Documents
A R T I C L E I N F O A B S T R A C T
Handling Editor: Prof. H. Timmermans Consumers prefer human service over chatbots in some contexts. We draw on treatments of trust in information
systems research to propose a three-tiered model of trust which posits that customer outcomes following chatbot
Keywords: service are influenced by dispositional trust in service technologies, generalized learned trust in chatbots, and
Customer service particularized situational trust in one’s chatbot service-giver. Our empirical tests, based on 643 participants in 3
Chatbots
studies, show that dispositional trust partially mediates the effect of learned trust. We also evaluate two alter
Artificial intelligence
native chatbot service configurations to help improve customer acceptance. The approach which combined
Trust
Dispositional trust chatbot service with a light touch human intervention matched the human service condition on most customer
Trust propensity outcomes.
Learned trust
Situational trust
Satisfaction
Loyalty
Recommendation acceptance
In an era dominated by technology, humans will always win on perform complex service tasks while using chatbots predominantly in
experience. That’s why Internova Travel Group, one of the industry’s supporting roles. The advent of generative AI programs, such as
largest travel services companies … announced the launch of a new ChatGPT and Bard, promises to finally enable full automation of
advertising campaign to remind consumers that choosing to book travel customer service encounters.
through humans may be the best travel decision they could ever make Chatbot-based customer service offers critical advantages over
(Internova Travel Group, 2021). human service, such as lower cost at scale, 24/7 on-demand availability,
Along with ease of use, trust is a key component of a consumer’s and scalability (i.e. Pizzi et al., 2021). A customer service chatbot
satisfaction with a digital experience (Gartner, Inc. 2018). platform adopted in a single organization can be made uniform with
respect to agent training and expertise, and be unaffected by factors that
1. Introduction normally degrade human performance, such as fatigue, memory lapses,
and distractions. Understandably, organizations have a powerful
A chatbot is a computer program designed to conduct human con incentive to replace human agents with artificial intelligence (AI)
versation via text online or via voice using text-to-speech (Ciechanowski chatbots in customer service (Wang et al., 2022; Sands et al., 2021).
et al., 2019; Mariani et al., 2023). Although chatbots have existed since To achieve success with full automation, organizations will first need
the 1960s, dramatic advances in artificial intelligence (AI) and natural to overcome consumer mistrust of chatbot service. Surveys of consumer
language processing are driving a recent wave of chatbot adoption in sentiment toward AI consistently show that consumers view chatbots as
commercial customer service applications (i.e., Aleedy et al., 2019; being less understanding, empathetic or flexible, and harder to deal with
Shumanov and Johnson, 2021). Yet, until recently, chatbots’ functional than human agents (e.g., Li et al., 2023; Hsu and Lin, 2023; Clark, 2022;
limitations have pressed organizations to rely on human agents to Gartner, 2018; Sheehan et al., 2020; Beaner, 2022; Pegasystems, 2019;
* Corresponding author. Maine Business School. 5723 DP Corbett Business Building, Orono, ME 04469.
E-mail addresses: dongling.huang@csun.edu (D. Huang), dmitri.markovitch@maine.edu (D.G. Markovitch), rusty.stough@maine.edu (R.A. Stough).
1
The authors contributed equally to this research. The names are listed alphabetically.
https://doi.org/10.1016/j.jretconser.2023.103600
Received 26 August 2023; Received in revised form 7 October 2023; Accepted 9 October 2023
Available online 17 October 2023
0969-6989/© 2023 Elsevier Ltd. All rights reserved.
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
Rese et al., 2020; Hill et al., 2015). The referenced studies all report tested extensively before launch (Bowman, 2023; Roose, 2023). A key
lower consumer trust in and preference for chatbot service, compared challenge is that such novel behaviors are hard to predict and correct
with human agents. Of note, the most capable generative AI chatbots because they are dynamic (e.g., see Bowman (2023) for an overview and
making the headlines are controversial for different reasons, and raise a Simon Willison’s weblog at https://simonwillison.net/2023/for current
host of other concerns involving their ethics, accuracy, and trustwor commentary on generative AI).
thiness (e.g., Bowman, 2023; Pegoraro et al., 2023; Roose, 2023; Jabeur Importantly, trust is known to affect perceptions of customer service
et al., 2023). provision, value, and outcomes as experienced or perceived. Trust in
Inspired by the practical challenge of achieving consumer accep fluences how consumers experience exchanges and interactions by
tance of chatbot service, we set out to explore the role of trust in how reducing perceived risk, anxiety, and discomfort (McLean et al., 2021;
consumers experience and respond to chatbots. We propose a novel Pavlou, 2003). Thus, higher trust in a provider can make service in
three-tiered model of trust in chatbots which posits that customer out teractions relatively more pleasant and satisfying. Trust motivates
comes following chatbot service provision and, hence, chatbot accep greater acceptance of provider recommendations, which, depending on
tance, are influenced by one’s propensity to trust service technologies, the service context, may enable better regimen compliance or more
generalized learned trust in chatbots, and particularized situational trust appropriate product selection and use (McKnight et al., 2002), thus
in one’s chatbot service-giver. This is our first contribution. Our second, enabling better results and user experience. Consumers who trust the
empirical, contribution consists in testing the impact of trust on con provider also show greater commitment and belief in the value of its
sumer chatbot acceptance. This contribution addresses a notable gap in offerings while spending less effort to consider alternative solutions
the literature (see Web Appendix A for a summary). Despite the prolif (Morgan and Hunt, 1994). Consistent with these results, empirical
eration of theoretical and empirical articles addressing the impact of AI research finds a positive relationship between trust and customer
on marketing phenomena, empirical research exploring the impact of satisfaction, including as a mediator variable, although the direction of
trust on customer experience with chatbot service has been quite causality has not been fully established (Chu et al., 2012; Geyskens et al.
limited. 1998).
We present three studies, collected online in the United States of
America, which test our research hypotheses. We use Study 1 (n = 199) 2.2. The trust construct
to confirm that individuals will report higher customer satisfaction,
repatronage intentions, and service agent’s recommendation acceptance Most researchers conceptualize trust based on their specific disci
(henceforth referred to as our focal customer outcomes) when served by plinary perspective, which has given rise to varying definitions of trust
a human agent versus a chatbot in our research setting. Study 1 also (McKnight et al., 2002). In customer service situations, customers can
explores whether an individual’s (1) propensity to trust service tech never be certain that the provider’s representative is knowledgeable and
nologies or (2) generalized learned trust in chatbots impact the focal genuinely willing to help. Hence, the marketing literature tends to
customer outcomes when individuals are served by a chatbot. Study 2 (n emphasize in its conception of trust the trustor’s confidence in the
= 225) primes trust in the chatbot service provider to test whether trustee’s competence and reliability. For example, Morgan and Hunt
manipulating provider-specific situational trust can impact customer (1994, p. 23) define trust as the feeling of “confidence in the exchange
experience with chatbot service. Study 3 (n = 219) evaluates two partner’s reliability and integrity.” Garbarino and Johnson (1999, p. 73)
chatbot service configurations informed by our trust framework to see if define trust in organizational settings as “customer confidence in the
they can match the human service condition on the focal customer quality and reliability of the services offered by the organization.”
outcomes. Barber (1983, p. 14) notes that trust generally entails “the expectation of
a technically competent role performance.”
2. Theory and hypotheses Various models of trust have been proposed in psychology, sociol
ogy, engineering, information systems, and the business disciplines.
2.1. The role of trust in consumer acceptance of chatbot customer service Most models identify the trustor (e.g., a consumer), the context, and the
trustee (e.g., a provider’s sales or service agent) as distinct sources of
Trust plays a central role in business and social transactions and variability in trust (Lee and See, 2004; Mayer et al., 1995). The mar
relationships (Dwyer et al., 1987). It influences consumers’ willingness keting literature, including service research, has, most extensively,
to engage in market exchanges and rely on providers (Garbarino and employed theoretical frameworks and measures centered on the trustee.
Johnson, 1999). Trust is a known factor in automation acceptance (Hoff This literature emphasizes the critical importance of the provider’s
and Bashir, 2015). Research in the “computers as social actors” para perceived competence, integrity, and benevolence in instilling trust (e.
digm shows that people react socially to computers, which also places g., Basso and Pizzutti, 2016; Doney and Cannon, 1997; also, see a
trust at the center of human-computer interactions (e.g., Nass and Moon, meta-analysis by Geyskens et al. 1998).
2000; Xu et al., 2022). Glikson and Woolley (2020, p. 230) note that Our theoretical development deviates from this paradigm in that it
“trust is particularly relevant to the human-AI relationship because of builds on trust frameworks developed in the information systems liter
the perceived risk embedded in human-AI relations, due to the ature and human factors research, which are both thematically related
complexity and non-determinism of AI behaviors.” In other words, to our exploration. The specific model we build on argues that the
consumers may never sufficiently understand AI and, hence, are trustor-, trustee-, and context-based factors find a distinct expression in
compelled to rely on some degree of trust when interacting with AI three inter-related trust layers: dispositional trust, learned trust, and
agents. situational trust (Hoff and Bashir, 2015; Mariani et al., 2023; Marsh and
Trust is particularly important in consumer acceptance of generative Dibben, 2003; McKnight et al., 2002). Dispositional trust encompasses a
AI chatbots as service-givers. These chatbots combine highly evolved trustor’s stable personality characteristic of being more or less trusting
capabilities that enable human-level performance with fundamental towards others. Learned trust reflects one’s trust behavior towards a
limitations that undermine their trustworthiness. Central to the particular entity arising from experience and other learning behaviors,
contradiction is the fact that generative AI uses machine learning to infer whereas situational trust reflects one’s trusting stance adjusted based on
information. This process is known to produce inaccuracies and outright situational cues present during an exchange (e.g.,Marsh and Dibben,
falsehoods, which, nevertheless, sound highly plausible. Equally 2003). Hence, the model presents a dynamic perspective on trust which
important, “as language models (LMs) scale, they develop many novel evolves in the moment and over time.
behaviors, good and bad” (Perez et al., 2022, p. 1). For example, the Nevertheless, the prevalent approaches to studying trust in tech
early version of Bing Chat displayed troubling behavior in spite of being nologies show a notable limitation. As McKnight et al. (2011, p. 2) point
2
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
out, even “within the Information Systems (IS) domain, as in other technology (Rau et al., 2009). However, longer-term, dispositional trust
fields, trust is usually examined and defined in terms of trust in people may evolve, influenced by accumulating experiences (Hoff and Bashir,
without regard for trust in the technology itself.” Therefore, we extend 2015), which initially inform learned trust. People also develop greater
the model to reflect the specificities of service interactions involving a trust towards other individuals and technologies as they age (e.g.,
non-human trustee. Our model proposes three layers of trust which, a Greiner and Zednik, 2019). For example, Ho et al. (2005) find that older
priori, are relevant to chatbot acceptance—(1) dispositional trust to individuals trust and rely more on technology-based decision aids than
wards service technologies, (2) learned trust in chatbots, in general, and younger people.
(3) situational trust in the specific chatbot providing a service. Of note, Importantly, the degree of one’s trust in a trustee directly influences
this distinct combination of trust factors reflects two ubiquitous service one’s experiences with the trustee, including satisfaction (Chu et al.,
scenarios—when a consumer interacts with a particular service pro 2012; Geyskens et al. 1998). “Trusting stance means that one assumes …
vider’s chatbot for the first time or only once. In either case, there is no better outcomes from dealing with people as though they are well
particularized learned trust in the chatbot representing the provider, but meaning and reliable” (McKnight et al., 2002, p. 340). These trust effects
all consumers will vary in terms of dispositional trust, and have prior also extend to human-AI interactions (for an overview of the literature
general experience involving chatbots or AI assistants and, hence, have on human reactions to AI, see Glikson and Woolley, 2020). Hence, in
developed more or less generalized learned trust in chatbots. dividuals with higher dispositional trust in technology are likely to
perceive their chatbot interactions in a more favorable light than in
2.3. Trust in people versus trust in technology dividuals who score low on dispositional trust in technology. Formally.
H1. Dispositional trust in service technologies is positively associated
There are important conceptual similarities between trust in people
with customer satisfaction, repatronage intentions, and recommenda
and trust in technology. Both trust phenomena arise when one chooses
tion acceptance following interactions with chatbot customer service.
to rely on another party (i.e., either a human or a chatbot) to achieve an
objective under uncertainty and lack of total control. Both phenomena
2.5. Learned trust
also involve expectations of competent performance, although, a chat
bot’s perceived competency will derive from its perceived adequacy for
The development of trust is a dynamic process (Lee and See, 2004).
the task (van Wezel et al., 2021). In both cases, there is an expectation of
In the absence of relevant information or prior experience, individuals
benevolence—that the agent’s purpose is to help, and also that human or
may adopt a trusting stance towards a service provider based on their
chatbot service agents will be honest with the customer (Chung et al.,
trust propensity. Subsequently, consumer trust towards a service pro
2020). Hence, chatbots can be evaluated based on the same fundamental
vider will evolve as consumers interact with the provider, experience
trust antecedents as human trustees. Moreover, the nomological
service outcomes, and update their knowledge structures, beliefs, and
network surrounding trust in technology is similar to trust in people and
attitudes (Mayer et al., 1995; McKnight et al., 2002). Hence, personal
includes the common trust types and linkages attributed to interpersonal
experience is a key mechanism in trust development, although, this trust
relationships (McKnight et al., 2011).
layer is also influenced by word of mouth (e.g., Meuter et al., 2013; Kong
et al., 2020), and various learning processes, including observational
2.4. Dispositional trust in service technologies
learning (Zhang, 2010). Accordingly, the resultant form of trust is often
referred to as learned trust. Where it exists, learned trust is believed to
Mayer et al., 1995, p. 715), in their widely cited work, define
supersede dispositional trust in its impacts on consumer behavior
dispositional trust as an individual’s “general willingness to trust
(Marsh and Dibben, 2003).
others.” Also called trust propensity, dispositional trust represents a
Learned trust initially develops in the form of particularized trust, or
personality trait that reflects an individual’s tendency to see the best in
trust towards a specific trustee (e.g., Glanville and Shi, 2020). Over time,
others and have a kind of belief in humanity that leads one to be more
as individuals interact with (or learn about) multiple entities in a subject
trusting irrespective of other factors, including the provider’s attributes
class, they develop trusting beliefs towards the entire class of similar
(McKnight et al., 1998).
trustees. An accumulation of positive experiences, hence, engenders
In a similar vein to interpersonal trust, individuals display variability
generalized learned trust towards the ‘average’ representative of the
in trust propensity towards technology in general (Hoff and Bashir,
class (e.g., Freitag and Traunmüller, 2009). Various studies in the field
2015), as well as distinct classes of technologies (McKnight et al., 2011).
of robotics, automation, and AI find a similar trajectory in human trust
In the latter case, a trusting stance towards a class of technologies rep
towards AI (Glikson and Woolley, 2020; Ullman and Malle, 2017). For
resents a lower-order construct, whereby trusting beliefs emanate from
instance, Bartneck et al. (2009) show that a positive interaction with a
one’s dispositional trust in technology (Serva et al., 2005). This pre
robotic pet improved the study participants’ attitude towards robots in
disposition to trust specific technologies essentially reflects beliefs that
general.
the technologies have beneficial attributes. For example, an individual
Based on this literature, we assume that learned trust in some service
with a higher propensity to trust in service technologies (e.g., automated
providers extends to similar provider types. In particular, a consumer
teller machines (ATMs), self-service kiosks, automated customer service,
who has experienced satisfactory chatbot-based customer service by
etc.) is more likely to assume that these technologies are reliable,
some providers is more likely to develop generalized trust in chatbots
functional, safe, useful, and help to improve one’s quality of life
and expect similarly satisfactory outcomes in future service encounters
(McKnight et al., 2011). The trust-in-automation research stream fea
involving other chatbots. Drawing on the research which shows a pos
tures several studies which explore dispositional trust in particular
itive relationship between trust and one’s experiences with a trustee
technologies. For instance, Biros et al. (2003) report that individuals
(McKnight et al., 2002), we posit that higher learned trust in chatbot
with a higher dispositional trust in computers showed greater accep
service, in general, will favorably impact customer outcomes following
tance of inputs from drones.
specific chatbot service encounters. Stated formally:
Unlike learned trust, which evolves following every interaction, and
situational trust, which is influenced by context in the moment, dispo H2. Generalized learned trust in chatbots is positively associated with
sitional trust tends to be relatively stable across situations, because it is a service satisfaction, repatronage intentions, and recommendation
feature of one’s psychological makeup. People with different life’s ex acceptance following interactions with chatbot customer service.
periences, personality traits and cultural backgrounds differ in their Moreover, scholars often model trust as a mediating variable which
dispositional trust. For example, the Japanese are known to be more links various antecedents with important outcomes, including those
trusting than others (e.g., Fukuyama, 1995), including towards robotic studied in this research (for an overview, see meta-analysis by Geyskens
3
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
et al. 1998). For example, Benamati et al. (2009) report that trust in an trustee or the service environment (e.g., Geyskens et al. 1998). Unlike
online vendor mediates consumer intentions to transact with its website. dispositional trust, which represents a durable personality trait, internal
In their study of supplier-retailer dyads, Morgan and Hunt (1994) find situational trust antecedents are more transitory and context-specific.
support for trust as a key mediator between relationship antecedents, One widely explored trust antecedent is domain expertise, or con
such as expected benefits and shared values, and relationship outcomes, sumer knowledge (Hoff and Bashir, 2015). Expertise has a profound
including acquiescence, loyalty, and cooperation. Auh (2005) extends impact on consumer behavior (Alba and Hutchinson, 1987), including
Morgan and Hunt’s key mediating variable model to show that trust the extent of information search and acceptance of external product
fully mediates the effect on loyalty of a service provider’s social and reviews. For example, Chinchanachokchai et al. (2021) show that
relational (i.e., soft) attributes, but only partially mediates the effect of domain experts have higher confidence in user-generated product rec
hard attributes, such as perceived performance quality. In research that ommendations than AI-generated recommendations. Other internal
is directly related to our investigation, Lin et al. (2021) examine con factors influencing situational trust include the trustor’s self-efficacy,
sumer acceptance of a virtual salesperson’s recommendation when it mood, and attention capacity during an interaction (Hoff and Bashir,
conflicts with other customers’ recommendations. They find that trust in 2015; Lee and See, 2004). For example, greater computer self-efficacy
a virtual salesperson fully mediates the link between conflicting reviews positively impacts trust in decision support systems (DSS), compli
and the willingness to follow a virtual agent’s recommendation. In ance, and DSS adoption (Madhavan and Phillips, 2010).
contrast, Ameen et al. (2021) report that consumer trust only partially External influences on situational trust are contextual cues that
mediates the perceived impact of AI-enabled service quality, conve impact consumer perceptions of a service provider’s trustworthiness.
nience, and personalization on AI-enabled customer experience. Hence, some of these factors pertain to the task and circumstances,
Consistent with this, the trust literature points to a likely causal including task difficulty, task framing, and risk perceptions (Hoff and
ordering between dispositional and learned trust. Dispositional trust is Bashir, 2015). Structural characteristics of the service environment also
typically treated as the highest-order construct. It is believed to influ play a notable role (McKnight et al., 2002. These are third-party
ence one’s behavior in risky unfamiliar contexts, including interactions mechanisms that motivate situational trust by minimizing the
with unfamiliar (types of) service providers and AI chatbots (Hasan perceived risk or otherwise enhancing the likelihood of the provider
et al., 2021; Schultz, 2006). Over time, the other party’s trustworthiness performing as expected. Examples of such “structural assurance” factors
is revealed. In this way, dispositional trust acts as an important mech include legal protections, protections from credit card companies, es
anism which facilitates experiential learning and learned trust forma crows, warranties, third party certifications, and seals of approval
tion. Subsequently, in situations involving familiar trustees, one’s (Burke, 2002; Kennedy et al., 2001).
behavior and outcomes will be influenced increasingly by learned trust. In addition, a service provider’s attributes may greatly influence
The impact of dispositional trust on outcomes will also increasingly flow perceptions of trustworthiness (e.g., Chen and Dhillon, 2003; Doney and
through learned trust. In other words, learned trust represents a Cannon, 1997). Several authors propose models which present con
second-order construct which likely mediates the effect of dispositional sumer trust as a function of perceived provider competence, benevo
trust. Formally: lence, integrity, and reliability (Mayer et al., 1995; McKnight et al.,
2002). Consumer perceptions of these attributes may develop over
H3a. Generalized learned trust in chatbots mediates the effect of
multiple interactions with the provider, and, hence, motivate learned
dispositional trust in service technologies on satisfaction, repatronage in
trust, as detailed in the preceding section. Or, they may be informed by
tentions, and recommendation acceptance following interactions with
new contextual cues and impact consumer trust during the encounter.
chatbot customer service.
The latter may include the provider’s apparent professionalism (Gregori
Our hypothesis H3 is motivated by the view of causal relationships
et al., 2014), brand recognition (Doney and Cannon, 1997), evidence of
between trust components which is prevalent in the trust literature (e.g.,
offline presence (Chen and Dhillon, 2003), or website characteristics,
Mayer et al., 1995; McKnight et al., 2011). However, alternative per
such as ease of use, usefulness, or even visible links to human customer
spectives also exist. There is a less well-developed, primarily theoretical,
support (Gefen, 2000). Multiple authors also find that provider
view which points to dispositional trust as a possible mediator of learned
communication, reputation, and evidence of past performance are
trust. First, there is evidence of reverse causality. One’s trust propensity
particularly germane to trust formation (e.g., Chen and Dhillon, 2003;
is relatively stable, but it is not fixed—changes simply occur over a
Geyskens et al. 1998). Importantly, the same provider attributes that
longer time-frame. They are influenced by accumulated experiences and
influence consumer trust in human agents also impact trust in AI tech
feedback concerning outcomes linked to trusting behaviors (Hoff and
nologies (Glikson and Woolley, 2020). In addition, technology-related
Bashir, 2015; McKnight et al., 2011). Additionally, technologies evolve,
attributes, such as disclosure of information about the operating rules
and dispositional trust will reflect updated beliefs that capture tech
governing the technology, may further affect situational trust in AI
nology advancement and learned trust experience. In other words,
(Glikson and Woolley, 2020; Hoff and Bashir, 2015).
learned trust also induces changes in dispositional trust. Moreover,
In sum, extant research indicates that presenting favorable infor
dispositional trust may continue to affect one’s trusting beliefs and be
mation about a service provider’s attributes, certifications, success rate,
haviors after trustworthiness has been established. It acts as a lens
and other measures of past performance may have a priming effect and
through which one sees social interactions and which is used to interpret
enhance consumers’ particularized situational trust in that provider.
trustees’ actions (Govier, 1994; Lewis and Weigert, 1985). We formalize
Higher trust, in turn, should result in better consumer evaluations of
this perspective on the role of dispositional trust as a competing hy
their service experiences with the provider. Hence, we test the following
pothesis as follows:
hypothesis:
H3b. Dispositional trust in service technologies mediates the effect of
H4. Individuals with higher particularized situational trust in a chatbot
generalized learned trust in chatbots on satisfaction, repatronage in
providing service will display higher satisfaction, repatronage in
tentions, and recommendation acceptance following interactions with
tentions, and recommendation acceptance following their service in
chatbot customer service.
teractions with the chatbot than individuals with lower particularized
situational trust.
2.6. Situational trust
3. Manipulating trust to improve customer outcomes
Situational trust is the trust that develops during a particular service
interaction in a particular setting. It is a function of multiple factors,
Since improving customer trust outcomes in chatbot service
some of which are internal to the trustor, and others—specific to the
4
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
encounters is an important motivation for this research, we next element with a lower-trust element in a system should additively in
consider the performance of two practical service configuration ap crease the total trust in the system (e.g., Amin-Naseri and Gilbert, 2014).
proaches suggested by our theoretical framework. The customer out The established positive relationship between trust and satisfaction (e.
comes attained by a human service agent represent a natural benchmark g., Chu et al., 2012) means that consumers will likely feel more positive
for this investigation. about the total service experience delivered by a chatbot-human agent
The preceding discussion indicates that situational trust towards a team.
specific provider can be influenced by contextual cues. This malleability The second mechanism whereby a human agent’s participation
makes situational trust an ideal focal point for a marketing action. As should boost trust in the chatbot component of the service team involves
hypothesized in H4, higher particularized situational trust is expected to the well-documented halo effect. It is a cognitive bias which entails the
be associated with generally better customer outcomes. This enables us tendency for positive impressions of a salient person or characteristic to
to set up our first “non-inferiority” performance test which we express in influence how other related persons or characteristics are judged (e.g.,
the following hypothesis:2 Nisbett and Wilson, 1977). The halo effect has been observed in a variety
of team and system contexts, which are, essentially, similar to our pro
H5. Following a service encounter, individuals with high particularized
posed “service team” or “service system” setting. For example, business
situational trust in their chatbot service-giver will report lower satisfac
scholars find that managers tend to provide more positive evaluations of
tion, repatronage intentions, and recommendation acceptance than in
subordinates who work in close proximity to them (Ferris et al., 1994).
dividuals in the human-service-agent condition.
In systems research, studies routinely find that multicomponent ratings
exhibit higher than expected correlation among the rated components.
3.1. Low-cost combination of chatbot and human customer service For example, movie ratings display an elevated correlation across the
typical movie attributes (Sahoo et al., 2012). Moreover, the halo effect
While fully automated customer service may be the ultimate goal for extends to perceived trustworthiness of technologies. For example,
most organizations, consumer mistrust of chatbot service in some con Gabrieli et al. (2021) report that placing images of trustworthy-looking
texts may motivate a gradual transition whereby chatbots will be com human faces on ATMs increased the perceived trustworthiness of the
bined with human agents in distinct ways to achieve operational or ATMs compared with the ATMs that featured less trustworthy-looking
marketing objectives.3 This projection is supported by practitioner sur faces.
veys which show that a higher proportion of consumers still prefer to be Based on these mechanisms, we proceed to evaluate the proposed
served by humans rather than chatbots even for basic tasks, such as chatbot-human service configuration on the focal performance measures
opening an account or making a purchase (e.g., Beaner, 2022; Chen relative to the human-only configuration. We are primarily interested in
et al., 2022). To the extent that the mistrust in chatbot service may be the chatbot-human configuration’s performance under conditions of low
deep-rooted and “sticky,” only manipulating particularized situational situational trust in the chatbot. However, it is also relevant to consider
trust may be insufficient to achieve satisfactory customer outcomes. We, the impact that adding a human agent may have when consumers have
hence, consider whether using a chatbot in the lead role with human high trust in the chatbot. We use the non-inferiority experimental design
support may engender higher consumer trust overall. to test the following set of hypotheses:
One practical chatbot-human service configuration that seemingly H6a. The proposed chatbot-human service configuration will under
satisfies efficiency objectives could feature a front-line chatbot agent perform human-only service on the focal customer outcomes under
augmented with a “light-touch” human intervention at the conclusion of conditions of low particularized situational trust in the service chatbot.
a service encounter. In this setup, the chatbot would be tasked with
resolving a customer issue or making a sale in its entirety. In situations H6b. The proposed chatbot-human service configuration will under
flagged as non-routine, the encounter would conclude by handing the perform human-only service on the focal customer outcomes under
consumer over to a human service specialist whose objective would be conditions of high particularized situational trust in the service chatbot.
to reassure the consumer and upsell, where appropriate. The human We summarize our hypotheses and key flows in Fig. 1. Note that the
intervention would be selective, last seconds in many cases, and, consumer preference factors across subjects are controlled through
therefore, cost little per service encounter on average while generating random assignment to our experimental conditions. The service quality
incremental sales revenue. element of the model is a major factor which affects customer outcomes.
There are two plausible mechanisms whereby even a fleeting human It is assumed to be constant across all our conditions.
agent’s addition to chatbot-led service may benefit customer outcomes.
First, consumers currently perceive human service agents as being 4. Research method
generally more trustworthy than chatbots (e.g., Beaner, 2022; Clark,
2022; Gartner, Inc. 2018; Pegasystems, 2019). Based on our framework, 4.1. Participants and Procedure
this engages higher dispositional and learned trust, since these trust
layers are interlinked through a feedback loop. Combining a higher-trust We conducted three experimental studies using the vignette method
(e.g., Jörling et al., 2019; Harborth and Pape, 2021) to test our research
hypotheses. We used Amazon’s MTurk platform to recruit study par
2
Non-inferiority experimental designs, which are prevalent in medicine, test ticipants and recruited participants from the United States in the fall of
whether a proposed treatment is at least as good as the standard-of-care. A 2022. We used a power analysis to ensure adequate samples. Partici
positive result can help establish that switching to the new treatment will do no pants were randomly assigned to six experimental conditions across the
harm. Our experimental challenge is essentially the same—we seek to identify three studies. All the studies opened with scales measuring (1) gener
service configurations that will at least match human service on key customer alized learned trust in chatbots and, separately, learned trust in human
outcomes. Since demonstrating non-inferiority relative to the human service service agents, and (2) dispositional trust in service technologies and,
benchmarks is the objective, we adopt the “non-inferiority” testing framework.
separately, dispositional trust in service personnel. We randomized the
This entails framing the null hypotheses so that the new treatment is posited to
order in which the question blocks about chatbots and human agents
be inferior to the benchmark. The alternative hypothesis is that the new
treatment is non-inferior. A non-inferiority claim is supported if the null hy were presented.
pothesis (of lower performance) is rejected. Next, we introduced our experimental vignette (see Web Appendix B
3
Although organizational adoption of generative AI is not the focus of this for vignettes and primes) which had the same opening paragraph in all
research, it is noteworthy that business leaders also cite trust as a barrier to AI three studies. It instructed the participants to imagine that they were
adoption, according to a recent IBM survey (Wiggers, 2023). reserving a long-planned African safari, accompanied by a renowned
5
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
Fig. 1. A model of the three layers of trust and customer service outcomes. The dashed arrow shows the hypothesized alternative mediation pathway. Feedback
effects are not shown for parsimony reasons. HM designates human-agent service. Hypotheses in bold font have received empirical support. * Hypotheses H5, H6a,
and H6b were stated as inferiority hypotheses. Only H6b received non-inferiority support. See Footnote 1 for details.
tour guide. Participants imagined having an initial interaction online (Stella)-human agent service combination with participants being
with a reservation agent who was, depending on the condition, either assigned either to high-trust or low-trust conditions. As in Study 2,
human (Amy Sullivan) or a chatbot (introduced as “Stella, your AI Stella’s recommendation of a replacement tour guide included the same
Virtual Representative”). Participants were then shown a mock-up of additional information about Stella’s track record and implications for
their imagined online chat with the reservation agent. In all the studies, the decision. The participants were then instructed to imagine that they
the participants were informed that their much-preferred tour guide was accepted the reservation agent’s recommendation and concluded the
unavailable. Instead, the reservation agent recommended another, reservation details, at which point Stella handed the customer over to
obscure, tour guide, who was new to the company. There was no public the company’s human customer relations manager Marisa Diaz. Marisa
information available about the replacement guide, except for the reassured the customer about the replacement tour guide and offered
reservation agent’s assurance that the guide was “experienced and well additional tour options which were declined. This concluded the reser
trained, like all our guides.” This introduced an element of uncertainty vation interaction in Study 3.
and risk—the customer was reserving the vacation of a lifetime at a high
cost, but not getting everything they were hoping to get. Moreover, a key 4.2. Screens and attention checks
element of the vacation experience was fundamentally uncertain and, at
a minimum, not as good as the preferred tour guide. From this point, the We took several steps to ensure data quality. We asked all partici
vignettes varied across the studies in the following manner. pants whether they had ever made a hotel or vacation reservation and
In Study 1, the participants interacted either with Amy Sullivan, a been on a tour (of any length) with a tour guide. We excluded those who
human agent, or Stella, the chatbot. The participants did not receive any had not, since they may have difficulty relating to our research scenario.
further information to facilitate their decision. They were instructed to We also used three comprehension and attention checks and a separate
imagine that they accepted the reservation agent’s guide recommen check to ensure that participants served by Stella understood that Stella
dation and concluded the reservation. This study additionally embedded was a chatbot. We excluded participants who failed any of these checks.
filler information in both conditions in order to create visual and length Using these filters and checks reduced our total sample by 89 (out of
equivalency with our experimental conditions in Studies 2 and 3. We 732) participants and resulted in sample size variation across the
used this study to test hypotheses 1–3. conditions.
In Study 2, which tested hypotheses 4 and 5, the participants inter
acted only with Stella, the chatbot, but were randomly assigned to high-
4.3. Measures
situational-trust or low-trust conditions. Stella’s recommendation of a
replacement tour guide included additional information about Stella’s
All the response alternatives in our rating scales discussed below are
track record and unpacked its implications for the reservation decision.
measured based on a seven-point scale. Table 1 shows the construction
In the high-trust condition, Stella’s track record showed a 95 percent
of our key variables, controls (described in the following sections), and
customer satisfaction rating with Stella’s recommendations. The
the associated reliability coefficients.
vignette further interpreted the information for the participants,
Dependent variables. To test customer satisfaction we first use a two-
restating that Stella almost always made good recommendations. In the
item measure, Satisfaction_Index, which captures (1) the extent to which
low-trust condition, Stella’s track record showed a 50 percent customer
participants were satisfied with the reservation interaction overall and
satisfaction rating with Stella’s recommendations. The vignette further
(2) satisfaction with the reservation agent specifically. We also measure
interpreted the information by stating that Stella made good recom
perceived service quality satisfaction using the distinct dimensions of
mendations about half of the time. Study 2 participants were then
the SERVPERF scale (Cronin and Taylor, 1994). Over the years, elements
instructed to imagine that they accepted the reservation agent’s
of SERVPERF have been extensively employed to evaluate service
recommendation and concluded the reservation.
quality satisfaction during distinct service encounters (e.g., Elliott,
Study 3, which tested hypotheses 6a and 6b, featured a chatbot
1994; Rodrigues et al., 2011). Following the convention, we adapt
6
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
7
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
(1985) involvement scale. Participants responded on a 1–7 scale how differences are significant at p < 0.1. We also compared the conditions
important, interesting, and exciting going on an African safari was to based on a separate check about the extent to which the participants
them. The three items were combined in an equally weighted index. were concerned about AI replacing human interaction and jobs, since
Finally, we use a short-form Evans and Revelle (2008) propensity to strong feelings in this regard could bias our results. On average, our
trust people scale to measure dispositional trust in people. Note that our participants displayed low concern about AI, with mean responses
primary interest in this paper is in the trust constructs directly affecting ranging between 3.9 and 4.3 out of seven across the cells.
chatbot perceptions, and we include the dispositional trust in people We also review the pairwise correlations between our trust di
variable only as a control. mensions. We note a moderately high correlation of 0.53–0.68 between
dispositional and learned trust variables in all our chatbot cells. This is
5. Results consistent with the hypothesized relationship between those trust di
mensions. In our five chatbot cells, we note a low and mostly insignifi
As the first step in our analyses, we tested for significant differences cant correlation between those two variables and the dispositional trust in
between experimental conditions on our control and demographic var people, which we use as a control. We further note a significant 0.42
iables (see Web Appendix C for descriptive statistics). None of the mean correlation between the more closely related constructs of dispositional
Table 2
Regressions of customer outcome variables on measures of dispositional trust in service technologies and generalized learned trust in chatbot service in the sample that
received chatbot service. These analyses are based on Study 1.
IVs/DVs = − 1 − 2 − 3 − 4 − 5 − 6 − 7 − 8
Sat_Index Serv_Efficiency Serv_Respons. Serv _Empathy Serv_Reliability Serv_Assurances Provider Loyalty Recomm_Accept
Panel A: test of H1
Disp. Trust in Service Tech 0.522*** 0.465*** 0.246** 0.559*** 0.406*** 0.617*** 0.658*** 0.426**
(0.08) (0.09) (0.08) (0.09) (0.09) (0.09) (0.10) (0.13)
Experience w/Chatbots 0.14 0.14 0.13 0.09 0.03 0.09 (0.03) (0.06)
(0.10) (0.12) (0.10) (0.11) (0.11) (0.10) (0.12) (0.16)
Experience w/Human CSRa 0.10 0.01 0.03 0.19 0.00 0.09 0.09 0.04
(0.10) (0.11) (0.09) (0.10) (0.10) (0.10) (0.12) (0.15)
Involvement 0.10 0.10 0.05 0.03 0.12 0.15 0.18 − 0.14
(0.08) (0.09) (0.07) (0.08) (0.08) (0.08) (0.09) (0.12)
Disp. Trust in People 0.01 0.12 0.12 (0.03) 0.06 0.02 0.02 0.02
(0.11) (0.12) (0.11) (0.12) (0.11) (0.11) (0.13) (0.17)
adj. R2 0.348 0.241 0.139 0.338 0.191 0.397 0.309 0.072
F 11.97 7.541 4.323 11.53 5.853 14.58 10.23 2.588
p 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.03
N 104 104 104 104 104 104 104 104
Panel B: test of H2
Learned Trust in Chatbots 0.481*** 0.486*** 0.239** 0.529*** 0.482*** 0.561*** 0.584*** 0.23
(0.09) (0.09) (0.08) (0.09) (0.08) (0.09) (0.11) (0.13)
Experience w/Chatbots 0.17 0.17 0.14 0.13 0.05 0.14 (0.02) (0.07)
(0.11) (0.12) (0.10) (0.11) (0.10) (0.11) (0.13) (0.17)
Experience w/Human CSRa 0.09 0.00 0.03 0.18 0.01 0.08 0.08 0.01
(0.10) (0.11) (0.09) (0.10) (0.09) (0.10) (0.12) (0.16)
Involvement 0.07 0.05 0.04 − 0.01 0.06 0.09 0.14 − 0.16
(0.08) (0.09) (0.08) (0.09) (0.08) (0.08) (0.10) (0.13)
Disp. Trust in People 0.10 0.19 0.18 0.07 0.16 0.15 0.09 − 0.01
(0.12) (0.13) (0.11) (0.12) (0.11) (0.12) (0.15) (0.18)
adj. R2 0.286 0.24 0.123 0.298 0.272 0.329 0.227 0.001
F 8.996 7.32 3.807 9.51 8.465 10.8 6.858 1.026
p 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.41
Nb 101 101 101 101 101 101 101 101
Panel C: Full Model
Disp. Trust in Service Tech 0.365** 0.262* 0.148 0.378** 0.144 0.444*** 0.449** 0.433*
(0.12) (0.13) (0.11) (0.12) (0.11) (0.12) (0.15) (0.19)
Learned Trust in Chatbots 0.239* 0.313* 0.141 0.279* 0.386*** 0.268* 0.287* − 0.0574
(0.11) (0.13) (0.11) (0.12) (0.11) (0.11) (0.14) (0.18)
Experience w/Chatbots 0.146 0.16 0.13 0.108 0.0446 0.111 − 0.0415 − 0.0968
(0.10) (0.12) (0.10) (0.11) (0.10) (0.10) (0.13) (0.16)
Experience w/Human CSRa 0.111 0.0146 0.0404 0.198 0.0216 0.101 0.0975 0.0318
(0.10) (0.11) (0.09) (0.10) (0.09) (0.10) (0.12) (0.15)
Involvement 0.087 0.0661 0.0483 0.00367 0.0704 0.114 0.16 − 0.14
(0.08) (0.09) (0.08) (0.08) (0.08) (0.08) (0.10) (0.13)
Disp. Trust in People 0.0553 0.163 0.163 0.0287 0.143 0.099 0.0399 − 0.0614
(0.11) (0.13) (0.11) (0.12) (0.11) (0.11) (0.14) (0.18)
adj. R2 0.347 0.263 0.13 0.357 0.276 0.413 0.291 0.045
F 9.843 6.96 3.485 10.25 7.368 12.71 7.835 1.794
p 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.11
Nb 101 101 101 101 101 101 101 101
8
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
trust in people and dispositional trust in service personnel in the human- paths for dispositional and learned trust. As the first step, we fitted a full
agent cell. Taken together, the observed statistics are consistent with model with both trust variables (shown in Panel C of Table 2). The
our arguments that dispositional trust in people, dispositional trust in resultant main effect coefficients on the dispositional trust variable
service technologies, and generalized learned trust in chatbots represent largely maintained their magnitude and significance, unlike the co
distinct constructs which capture different dimensions of trust. efficients on the learned trust variable, which declined in magnitude and
We used ordinary least squares regression (OLS) to test our hy lost significance in some cases. Next, to test our hypotheses, we ran two
potheses. In addition to the relevant independent variables, all the re sets of mediation analyses for each dependent variable (summarized in
gressions included four control variables—(the extent of) experience Table 3). We used Hayes (2018) PROCESS macro with 5000 bootstrap
with chatbots, experience with human customer service, involvement samples. In the first analysis, we specified dispositional trust in service
with respect to African safari vacations, and dispositional trust in technologies as the independent variable and generalized learned trust in
people. chatbots as the mediator. This analysis produced only two significant (at
p < 0.05) coefficients on the indirect effect of dispositional trust, in the
5.1. Study 1 regressions of Service_Efficiency and Service_Empathy. Thus, our data
provide only limited support for H3a. We then evaluated the reverse
The first aim of Study 1 was to establish the baseline preference for mediation pathway with learned trust as the independent variable and
human or chatbot service in our research sample so that we could dispositional trust as the mediator. All but one of the resultant co
evaluate actions that can make chatbot service more attractive to con efficients on the indirect effect were positive and statistically significant
sumers. We fitted the regressions in the combined sample which (at p < 0.05), consistent with H3b. We conclude that, on balance, these
included human-agent and chatbot service conditions with the total of results provide stronger support for H3b than H3a. Additionally, we note
199 observations. The focal variable was the chatbot service-giver that we observe only partial mediation, since the direct path on learned
condition dummy. trust remains significant in six out of our eight regressions (see Panel C of
The coefficients on the dummy variable indicating chatbot service Table 2).
were significantly negative in all the regressions, and the p-values were
below 0.05 in all but one case. Concerning the latter, the coefficient on 5.2. Study 2
the service-giver dummy was not significant in the regression of Serv
ice_Efficiency (from the SERVPERF construct). While contradictory, this In Study 2, we primed participants for higher or lower situational
result makes intuitive sense, since the chatbots’ defining characteristic trust in the chatbot service-provider. Our purpose was to explore
at this time is their efficiency. Therefore, we conclude that our subjects, whether one’s level of particularized situational trust during a service
on average, strongly prefer human service. encounter impacts customer satisfaction with chatbot service and sub
We next explored whether one’s propensity to trust service tech sequent trusting behaviors. This study used 219 valid observations. As a
nologies and learned trust in chatbots, based on prior interactions, had a manipulation check, we asked participants to rate Stella’s trustworthi
positive effect on the focal outcomes, as posited in H2 and H3, respec ness in making tour guide recommendations. The manipulation check
tively. To test these hypotheses, we ran separate regressions with the means were 5.23 and 3.88 out of seven in the high-trust and low-trust
focal independent variables and controls in the subset served by Stella, conditions, respectively (p < 0.001), indicating that our priming had
the chatbot. We show the regressions results in Panels A and B of successfully created experimental conditions with different levels of
Table 2. We find the effect of dispositional trust in service technologies on situational trust. We next separately regressed our eight dependent
all the dependent variables to be significantly positive at p < 0.01. variables on a dummy variable for situational trust and the controls. The
Our results for the effect of generalized learned trust in chatbots are situational trust variable took the value of one in the high trust condition
largely similar, except in the Reservation_Acceptance regression, which is and zero otherwise. The results are shown in Tables 4 and 5. All the
not statistically significant, implying that dispositional trust may play a regressions are highly significant, with the coefficient on the High_
relatively more fundamental role in behavioral responses. We explore Situational_Trust dummy being positive and statistically significant (p <
this supposition further in our investigation of mediation paths detailed 0.01 or lower) in all the regressions. We conclude that our data provide
below. We also conclude that, taken together, our findings support H1 strong support for our hypothesis H4.
and H2.
Our hypotheses H3a and H3b specify two alternative mediation
Table 3
Summary of mediation analysis tests based on Hayes (2018) PROCESS macro with 5000 bootstrap samples. These analyses are based on Study 1.
Independent Variable Mediator Dependent Variable b 95% CI Significance
9
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
Table 4
Regressions of customer outcome variables on the (high or low) level of situational trust in the chatbot service-provider. These analyses are based on Study 2.
IVs/DVs = − 1 − 2 − 3 − 4 − 5 − 6 − 7 − 8
Sat_Index Serv_Efficiency Serv_Respons. Serv _Empathy Serv_Reliability Serv_Assurances Provider Loyalty Recomm_Accept
5.3. Study 3 the value of one if the reservation was handled by a human agent or
zero—by a chatbot-human service team.
Finally, Study 3 evaluated two approaches to stimulating trust in Looking first at the results based on the sample that was primed for
chatbots in order to achieve customer outcome parity with human ser low trust in Stella, we note that all the regressions are statistically sig
vice. The first approach entailed priming situational trust in the chatbot nificant. However, the coefficient on the HM_Service dummy is signifi
providing service. To test our hypothesis, we regressed our dependent cantly positive in every regression (at p < 0.05 and lower). Thus, all the
variables on the dummy variable labeled, HM_Service, which took the evidence strongly supports the “inferiority” hypothesis H6a. Stated
value of one for the human-agent service condition and zero for chatbot differently, we find no evidence that a brief human intervention
service. We show the results in Panel A of Table 5. While the coefficient following chatbot service can meaningfully improve customer outcomes
on the focal dummy was not statistically significant in four regressions, when consumers have low situational trust in the chatbot service-
it was significantly positive (at p < 0.05) in four other regressions provider.
involving Satisfaction_Index, Service_Empathy, Service_Assurances, and Next considering the sample with high situational trust in the chatbot
Repatronage. We note that this pattern of results makes intuitive sense, in Stella, we note that the coefficient on the HM_Service dummy is signifi
particular, the lack of impact on perceived empathy. On balance, our cantly positive only in the regression of Service_Empathy, meaning that
data provide partial support for H5, which is framed as a non-inferiority human service outperformed the chatbot-human agent combination
test. It posits that even when primed for high situational trust in the only on that customer outcome. The dummy coefficient in the other
chatbot service giver, individuals report worse outcomes compared with seven regressions is not statistically significant at p < 0.05. Although, we
equivalent service by a human agent. We find this to be true with respect note that the focal coefficient is marginally significant at p < 0.1 and
to half of our dependent variables. Reframing our results in a positive negative in the regression of Service_Efficiency, suggesting that the
light, one can, alternatively, state that priming particularized situational chatbot-human agent combination borderline outperformed the human
trust in the chatbot service-giver helped to achieve non-inferiority with service condition in terms of perceived service efficiency. These two
human service on half of the outcome measures. results align with the basic intuition that human agents will tend to score
We next evaluated our second proposed approach to stimulating higher on empathy in most circumstances, whereas chatbots are
trust in chatbot service. This approach involved comparing human ser generally associated with service efficiency. In sum, our data lead us to
vice with a chatbot-human combination primed for low or high situa reject the inferiority hypotheses H6b for seven out of our eight depen
tional trust. We estimated our regressions in two samples which dent variables. Stated differently, these results largely support our
contained 212 and 203 valid observations, respectively. In both samples, proposition that a chatbot-human-agent combination, as described in
we regressed our battery of dependent variables on a dummy which took this study, can match human-agent service on customer outcomes
10
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
Table 5
Regressions of the customer outcome variables on the service provider type—human agent vs. chatbot/human agent—in the subsets with low or high situational trust
in the chatbot. These analyses are based on Study 3.
IVs/DVs = − 1 − 2 − 3 − 4 − 5 − 6 − 7 − 8
Sat_Index Serv_Efficiency Serv_Respons. Serv _Empathy Serv_Reliability Serv_Assurances Provider Loyalty Recomm_Accept
provided the customers have situational trust in the chatbot. from satisfied customers. Moreover, a reputation for superior customer
service can bestow a competitive advantage, as exemplified by com
6. Discussion, managerial implications, and conclusions panies like Nordstrom’s and Apple (Garg, 2020). Hence,
customer-oriented organizations will likely seek ways to implement
Technology analysts proclaim that chatbots are rapidly becoming “a chatbots, as well as supporting structures and activities, in a manner that
critical technology component of a service organization’s self-service can help close the satisfaction gap between human and automated
strategy” (Gartner, Inc. 2023). At the same time, industry surveys customer service.
reveal that consumers generally prefer service by human representa We draw on our trust framework to propose two chatbot service
tives. We confirm this preference in our research setting. Compared with configurations that may stimulate customer trust during a service
equivalent service provided by a chatbot, our research participants re encounter and, thereby, help improve consumer experience with chat
ported higher satisfaction with human customer service. They also bot service. We empirically evaluate both approaches on their ability to
demonstrated greater provider repatronage intent and recommendation match human service on customer satisfaction, service agent’s recom
acceptance after a service encounter with a human agent. mendation acceptance, and repatronage intentions. The first approach
The tension between service providers’ chatbot adoption and con entails priming situational trust in the chatbot service provider at the
sumer hesitancy in accepting chatbot service presents a challenge for the point-of-service. We find this approach to be more effective than basic
adopting organizations in the long run. Many companies and non-profits chatbot service (without priming), although, falling short of human
strive for positive customer outcomes and marketing benefits which flow service on half of the customer outcome measures used in our tests. Our
11
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
Our experimental results corroborate industry surveys which point to 6.2. Theoretical implications
low consumer trust in chatbots as an impediment to chatbots’ wide
spread consumer acceptance and continued use. The development of AI This current research builds on previous research on trust in service
chatbots’ problem-solving and relational capabilities is likely to provide technologies (i.e. Mayer et al., 1995; Glikson and Woolley, 2020). It is
an effective solution to the trust conundrum in the long run. In the unique in that it combines theoretical lenses from marketing (i.e. Auh,
interim, we contend that marketing interventions, such as trust priming, 2005), information systems (i.e. McKnight et al., 2011) and human
and efficient chatbot-human service combinations may provide an factors research (i.e. Hoff and Bashir, 2015), which allows us to develop
improvement on service configurations that solely feature live or chat new approaches and insights. We propose a novel theoretical framework
bot agents. focusing on the role of trust in AI service encounters. We use the
To begin with, consumers vary in their preference for chatbot or framework to generate empirical evidence showing different outcomes
human service, as well as their dispositional and learned trust in chat involving consumer response to chatbot service based on the interplay of
bots. Hence, segmentation- and targeting-based approaches may present three distinct layers of trust: dispositional, learned and situational (i.e.
a low-hanging fruit on the path to improving chatbot-driven customer Marsh and Dibben, 2003). Hence, our research advances the trust
outcomes. For example, companies may leverage self-segmentation by literature by proposing a model of trust that is specifically adapted to
offering consumers a choice of chatbot or human service provider at the consumer interactions with chatbots. Another important theoretical
point-of-service (possibly, with the option of changing at any time be distinction of our framework is in presenting a dynamic view of trust
tween the two, to facilitate chatbot acceptance). Further self-targeting which accounts for changes in consumer trust over time. This opens new
may be achieved by enabling different, even a-la-carte, configurations avenues for research into temporal aspects of trust, in contrast to the
within chatbot service offerings (e.g., different chatbot personas). Firms prevalent approaches to trust that focus on trustee attributes, such as
that use chatbot service can also consider adding the trust variables as confidence, integrity, reliability, and benevolence.
additional inputs into their segmentation analysis. Further, our research provides insights into the process of trust for
Another promising chatbot implementation approach, which may be mation and elucidates the relationships between the focal trust layers. At
combined with self-segmented offerings, includes taking steps to stim the most fundamental level, one’s interactions with chatbots are shaped
ulate particularized situational trust in the chatbot service-giver. This by the extent of one’s propensity to trust service technologies—an
would include careful orchestration of contextual cues that can enhance assumption that these technologies are largely benign, beneficial, and
consumer confidence in the chatbot’s abilities (e.g., based on its success reliable. Over time, positive experiences with customer service chatbots
rate), honesty and benevolence. Conceptually and ethically, this would and other learning behaviors stimulate the development of generalized
not be different from the existing practice of human service providers learned trust in chatbots (i.e. Kong et al., 2020). In addition, during
displaying in their offices and on their marketing materials their degree service encounters with a chatbot, consumers may be exposed to
diplomas, testimonials, and prominent endorsements. We note that, contextual cues highlighting the chatbot’s capabilities, track record, or
while this approach was not completely on a par with human service in performance guarantees (i.e. Chinchanachokchai et al., 2021). These
our tests, it is very low-costand can be further enhanced using sophis cues motivate particularized situational trust in the specific chatbot
ticated priming instruments. service-giver.
However, the most promising and best-performing chatbot configu Our empirical investigation confirms that these trust layers impact
ration in our tests involved combining a chatbot in the primary service consumer experience with chatbots. To our knowledge we are the first to
role with a human agent in a supporting role. This approach naturally provide empirical evidence that all three trust layers have a direct effect
instills higher trust in the service team, although it performs better when on customer satisfaction with chatbot service and subsequent trusting
combined with a prime to enhance situational trust in the chatbot. To behaviors, including recommendation acceptance and repatronage in
the extent that a human agent can step in following the conclusion of a tentions. Dispositional trust in service technologies, in particular, was
service encounter, for example, for quality control and reassurance, the found to play a central role by mediating the effect of learned trust in
operating cost of such configuration may be only marginally higher than chatbots. This result conforms with the theoretical prediction that “ob
that of fully automated service. To reduce costs further, it can be servations are theory-laden” (Govier, 1994, p. 244) and trust propensity
selectively applied to non-routine service encounters. continues to impact one’s trusting behavior even in the presence of
We believe that the outlined chatbot-human implementation with a established trust.
situational trust prime represents a reasonable path towards both more
efficient and effective customer service that has many of the advantages 6.3. Limitations and future research
of chatbot and human agents, while requiring acceptable trade-offs.
Although it may be somewhat more expensive than full automation, Our research includes some limitations. First, while we designed our
and poses scaling challenges to incorporate the human element, it studies in such a way to limit potential confounds, we acknowledge that
promises superior customer satisfaction and important marketing ben there are additional variables which may influence the perception of the
efits, including incremental sales. If nothing else, the proposed config chatbot. For example, although we collected some demographic vari
uration may serve as a functional bridge towards full automation in the ables in our studies (i.e., gender and age), and verified that our results
future when consumers can be served by virtual agents endowed with are robust to their inclusion into the regression models, we did not have
high general and emotional intelligence (Prentice, 2019; Prentice et al., information on other additional socio-demographic variables such as
12
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
income, education and occupation. Culture is another dimension for Basso, K., Pizzutti, C., 2016. Trust recovery following a double deviation. J. Serv. Res. 19
(2), 209–223.
which we did not have control. Given that culture is not a typical control
Beaner, L., 2022. Two Thirds of Americans Feel Secure about Their Purchases when
variable in the literature, we leave it to further investigation on specific Dealing with a Live Person. SWNS. URL: https://swnsdigital.com/us/2022/10/t
service contexts. wo-thirds-of-americans-feel-secure-about-their-purchases-when-dealing-with-a-live-
Additionally, the demographics (race, age and gender) of the tour person/. Accessed on 5/9/2023.
Bowman, S.R., 2023. Eight Things to Know about Large Language Models. arXiv preprint
guides and service providers may have influenced the participants. Since arXiv:2304.00612.
the tour guide pictures and the chatbot avatar were the same across Benamati, J., Fuller, M.A., Serva, M.A., Baroudi, J., 2009. Clarifying the integration of
conditions (Web Appendix B), it is unlikely that any effects would in trust and TAM in e-commerce environments: implications for systems design and
management. IEEE Trans. Eng. Manag. 57 (3), 380–393.
fluence the results. However, future research should consider comparing Biros, D.P., Fields, G., Gunsch, G., 2003. The effect of external safeguards on human-
different demographics between conditions, and consider the possibility information system trust in an information warfare environment. In: Proceedings of
of gender influence. the 36th Annual Hawaii International Conference on System Sciences. IEEE,
pp. 1–10.
Next, consumers sometimes have a choice between using a chatbot Burke, R.R., 2002. Technology and the customer interface: what consumers want in the
immediately or queueing up for a human agent. This was not an option physical and virtual store. J. Acad. Market. Sci. 30 (4), 411–432.
in our study. We believe our setting is commonplace in this regard. At Chen, S.C., Dhillon, G.S., 2003. Interpreting dimensions of consumer trust in e-
commerce. Inf. Technol. Manag. 4 (2/3), 303–318.
the same time, we recognize that enabling the choice of a service-giver Chen, Q., Gong, Y., Lu, Y., Tang, J., 2022. Classifying and measuring the service quality
may affect one’s subsequent experiences. Since that various chatbot of AI chatbot in frontline service. J. Bus. Res. 145, 552–568.
service options are possible, as discussed in the preceding section, it Chinchanachokchai, S., Thontirawong, P., Chinchanachokchai, P., 2021. A tale of two
recommender systems: the moderating role of consumer expertise on artificial
would be instructive for future research to examine consumer trust
intelligence based product recommendations. J. Retailing Consum. Serv. 61,
patterns and related experiences in light of the distinct paths that con 102528.
sumers may take leading up to their service encounter. Chu, P., Lee, G., Chao, Y., 2012. Service quality, customer satisfaction, customer trust,
In the same vein, we contextualize this research in diverse customer and loyalty in an e-banking context. SBP (Soc. Behav. Pers.) 40 (8), 1271–1284.
Chung, M., Ko, E., Joung, H., Kim, S.J., 2020. Chatbot e-service and customer satisfaction
service applications that entail service provided by human and chatbot regarding luxury brands. J. Bus. Res. 117, 587–595.
agents via online chat. The degree to which our recommendations may Ciechanowski, L., Przegalinska, A., Magnuski, M., Gloor, P., 2019. In the shades of the
extend to face-to-face service contexts (e.g., involving humanoid ro uncanny valley: an experimental study of human–chatbot interaction. Future
Generat. Comput. Syst. 92, 539–548.
bots), or voice-based interactions, which, both, enable a distinct set of Clark, S., 2022. Do your customers trust your AI? Downloaded from. https://www.cmswi
cues and limitations would need to be established in future research. re.com/customer-experience/do-your-customers-trust-your-ai/. Accessed on 01/10/
Another future research direction would be to explore the boundary 2023.
Cronin Jr., J.J., Taylor, S.A., 1994. SERVPERF versus SERVQUAL: reconciling
conditions for chatbot service acceptance. For example, we pursue our performance-based and perceptions-minus-expectations measurement of service
research questions in a rather general setting that embeds moderate quality. J. Market. 58 (1), 125–131.
level of risk, and focuses on a mostly positive consumer experi Doney, P.M., Cannon, J.P., 1997. An examination of the nature of trust in buyer–seller
relationships. J. Market. 61 (1), 35–51.
ence—making a travel reservation. However, consumers often contact Dwyer, F.R., Schurr, P.H., Oh, S., 1987. Developing buyer-seller relationships. J. Market.
customer service with a problem (e.g., a missed flight, a product mal 51 (2), 11–27.
function, a return, a missing package, etc.). While our framework and Elliott, K.M., 1994. SERVPERF versus SERVQUAL: a marketing management dilemma
when assessing service quality. J. Market. Manag. 4 (2), 56–61.
recommendations generalize to chatbot service experiences in such
Evans, A.M., Revelle, W., 2008. Survey and behavioral measurements of interpersonal
settings, further empirical research in distinct settings is warranted. trust. J. Res. Pers. 42 (6), 1585–1593.
Ferris, G.R., Judge, T.A., Rowland, K.M., Fitzgibbons, D.E., 1994. Subordinate influence
and the performance evaluation process: test of a model. Organ. Behav. Hum. Decis.
Declaration of competing interest
Process. 58 (1), 101–135.
Freitag, M., Traunmüller, R., 2009. Spheres of trust: an empirical analysis of the
The authors declare that they have no known competing financial foundations of particularised and generalised trust. Eur. J. Polit. Res. 48 (6),
782–803.
interests or personal relationships that could have appeared to influence
Fukuyama, F., 1995. Dust: the Social Virtues and the Creation of Prosperity. Hamish
the work reported in this paper. Hamilton.
Gabrieli, G., Ng, S., Esposito, G., 2021. Hacking trust: the presence of faces on automated
Data availability teller machines (ATMs) affects trustworthiness. Behav. Sci. 11 (6), 91.
Garbarino, E., Johnson, M.S., 1999. The different roles of satisfaction, trust, and
commitment in customer relationships. J. Market. 63 (2), 70–87.
Data will be made available on request. Garg, P., 2020. How has Apple gained competitive edge over the years? Downloaded
from. https://medium.com/macoclock/how-has-apple-gained-competitive-edge-ove
r-the-years-aff0a259de0d. Accessed on 01/02/2023.
Appendix A. Supplementary data Gartner, Inc., 2018. Gartner Survey Finds that Most Consumers Have Underwhelming
Digital Experience. Available from: https://www.gartner.com/en/newsroo
Supplementary data to this article can be found online at https://doi. m/press-releases/2018-10-03-gartner-survey-finds-that-most-consumers-have-under
whelming-digital-experiences. Accessed on 10/14/2022.
org/10.1016/j.jretconser.2023.103600. Gartner, 2023. The Customer Service Chatbot Deployment Guide. Inc.. https://www.gart
ner.com/en/customer-service-support/trends/customer-service-chatbot-guide.
References Accessed on 01/01/2023.
Gefen, D., 2000. E-commerce: the role of familiarity and trust. Omega 28 (5), 725–737.
Geyskens, I., Steenkamp, J.B.E., Kumar, N., 1998. Generalizations about trust in
Alba, J.W., Hutchinson, J.W., 1987. Dimensions of consumer expertise. J. Consum. Res.
marketing channel relationships using meta-analysis. Int. J. Res. Market. 15 (3),
13 (4), 411–454.
223–248.
Aleedy, M., Shaiba, H., Bezbradica, M., 2019. Generating and analyzing chatbot
Glanville, J.L., Shi, Q., 2020. The Extension of particularized trust to generalized and
responses using natural language processing. Int. J. Adv. Comput. Sci. Appl. 10 (9),
out-group trust: the constraining role of collectivism. Soc. Forces 98 (4), 1801–1828.
60–68.
Glikson, E., Woolley, A.W., 2020. Human trust in artificial intelligence: review of
Ameen, N., Tarhini, A., Reppel, A., Anand, A., 2021. Customer experiences in the age of
empirical research. Acad. Manag. Ann. 14 (2), 627–660.
artificial intelligence. Comput. Hum. Behav. 114, 106548.
Govier, T., 1994. Is it a jungle out there? Trust, distrust and the construction of social
Amin-Naseri, M., Gilbert, S., 2014. A system dynamics approach to building team trust
reality. Dialogue 33, 237–252.
models: exploring the challenges. In: Pedagogy that Makes A Difference: Exploring
Gregori, N., Daniele, R., Altinay, L., 2014. Affiliate marketing in tourism: determinants of
Domain-independent Principles across. Instructional Management Research within
consumer trust. J. Trav. Res. 53 (2), 196–210.
the ITS Community, p. 49.
Harborth, D., Pape, S., 2021. Investigating privacy concerns related to mobile augmented
Auh, S., 2005. The effects of soft and hard service attributes on loyalty: the mediating
reality Apps–A vignette based online experiment. Comput. Hum. Behav. 122,
role of trust. J. Serv. Market. 19 (2), 81–92.
106833.
Barber, B., 1983. The Logic and Limits of Trust. Rutgers University Press.
Hasan, R., Shams, R., Rahman, M., 2021. Consumer trust and perceived risk for voice-
Bartneck, C., Kulić, D., Croft, E., Zoghbi, S., 2009. Measurement instruments for the
controlled artificial intelligence: the case of Siri. J. Bus. Res. 131, 591–597.
anthropomorphism, animacy, likeability, perceived intelligence, and perceived
safety of robots. International Journal of Social Robotics 1, 71–81.
13
D. Huang et al. Journal of Retailing and Consumer Services 76 (2024) 103600
Hayes, A.F., 2018. PROCESS: A Versatile Computational Tool for Observed Variable Parasuraman, A., Zeithaml, V.A., Berry, L.L., 1985. A conceptual model of service quality
Mediation, Moderation, and Conditional Process Modeling. and its implications. J. Market. 49 (4), 41–50.
Hill, J., Ford, W.R., Farreras, I.G., 2015. Real conversations with artificial intelligence: a Pavlou, P.A., 2003. Consumer acceptance of electronic commerce: integrating trust and
comparison between human–human online conversations and human–chatbot risk with the technology acceptance model. Int. J. Electron. Commer. 7 (3), 101–134.
conversations. Comput. Hum. Behav. 49, 245–250. Pegasystems, 2019. What Consumers Really Think about AI: A Global Study. Retrieved
Ho, G., Wheatley, D., Scialfa, C.T., 2005. Age differences in trust and reliance of a from. https://www.ciosummits.com/what-consumers-really-think-about-ai.pdf.
medication management system. Interact. Comput. 17, 690–710. Pegoraro, A., Kumari, K., Fereidooni, H., Sadeghi, A.R., 2023. To ChatGPT, or Not to
Hoff, K.A., Bashir, M., 2015. Trust in automation: integrating empirical evidence on ChatGPT: that Is the Question!. arXiv preprint arXiv:2304.01487.
factors that influence trust. Hum. Factors 57 (3), 407–434. Perez, E., Ringer, S., Lukošiūtė, K., Nguyen, K., Chen, E., Heiner, S., et al., 2022.
Hsu, C.L., Lin, J.C.C., 2023. Understanding the user satisfaction and loyalty of customer Discovering Language Model Behaviors with Model-Written Evaluations. arXiv
service chatbots. J. Retailing Consum. Serv. 71, 103211. preprint arXiv:2212.09251.
Internova Travel Group, 2021. The Future of Travel Is Human. Downloaded from. htt Pizzi, G., Scarpi, D., Pantano, E., 2021. Artificial intelligence and the new forms of
ps://internova.com/the-future-of-travel-is-human/. Accessed on 5/3/2023. interaction: who has the control when interacting with a chatbot? J. Bus. Res. 129,
Jabeur, S.B., Ballouk, H., Arfi, W.B., Sahut, J.M., 2023. Artificial intelligence applications 878–890.
in fake review detection: bibliometric analysis and future avenues for research. Prentice, C., 2019. Managing service encounters with emotional intelligence. J. Retailing
J. Bus. Res. 158, 113631. Consum. Serv. 51, 344–351.
Jacoby, J., Chesnut, R.W., 1978. Brand Loyalty: Measurement and Management. John Prentice, C., Wang, X., Loureiro, S.M.C., 2019. The influence of brand experience and
Wiley & Sons, New York, NY. service quality on customer engagement. J. Retailing Consum. Serv. 50, 50–59.
Jörling, M., Böhm, R., Paluch, S., 2019. Service robots: Drivers of perceived Rau, P.L., Li, Y., Li, D., 2009. Effects of communication style and culture on ability to
responsibility for service outcomes. J. Service Res. 22 (4), 404–420. accept recommendations from robots. Comput. Hum. Behav. 25, 587–595.
Kennedy, M.S., Ferrell, L.K., LeClair, D.T., 2001. Consumers’ trust of salesperson and Rese, A., Ganster, L., Baier, D., 2020. Chatbots in retailers’ customer communication:
manufacturer: an empirical study. J. Bus. Res. 51 (1), 73–86. how to measure their acceptance? J. Retailing Consum. Serv. 56, 102176.
Kong, Y., Wang, Y., Hajli, S., Featherman, M., 2020. In sharing economy we trust: Roose, K., 2023. A Conversation with Bing’s Chatbot Left Me Deeply Unsettled. New York
examining the effect of social and technical enablers on millennials’ trust in sharing Times. Retrieved from. https://www.nytimes.com/2023/02/16/technology/bing-c
commerce. Comput. Hum. Behav. 108, 105993. hatbot-microsoft-chatgpt.html.
Lee, J.D., See, K.A., 2004. Trust in automation: designing for appropriate reliance. Hum. Rodrigues, L.L., Barkur, G., Varambally, K.V.M., Motlagh, F.G., 2011. Comparison of
Factors 46 (1), 50–80. SERVQUAL and SERVPERF metrics: an empirical study. The TQM Journal 23 (6),
Lewis, J.D., Weigert, A., 1985. Trust as a social reality. Soc. Forces 63, 967–985. 629–643.
Li, S., Peluso, A.M., Duan, J., 2023. Why do we prefer humans to artificial intelligence in Sahoo, N., Krishnan, R., Duncan, G., Callan, J., 2012. Research note—the Halo effect in
telemarketing? A mind perception explanation. J. Retailing Consum. Serv. 70, multicomponent ratings and its implications for recommender systems: the case of
103139. yahoo! Movies. Inf. Syst. Res. 23 (1), 231–246.
Lin, Y.T., Doong, H.S., Eisingerich, A.B., 2021. Avatar design of virtual salespeople: Sands, S., Ferraro, C., Campbell, C., Tsao, H.Y., 2021. Managing the human–chatbot
mitigation of recommendation conflicts. J. Serv. Res. 24 (1), 141–159. divide: how service scripts influence service experience. J. Serv. Manag. 32 (2),
Madhavan, P., Phillips, R.R., 2010. Effects of computer self-efficacy and system 246–264.
reliability on user interaction with decision support systems. Comput. Hum. Behav. Serva, M.A., Benamati, J.S., Fuller, M.A., 2005. Trustworthiness in B2C e-commerce: an
26, 199–204. examination of alternative models. ACM SIGMIS - Data Base: the DATABASE for
Mariani, M.M., Hashemi, N., Wirtz, J., 2023. Artificial intelligence empowered Advances in Information Systems 36 (3), 89–108.
conversational agents: a systematic literature review and research agenda. J. Bus. Sheehan, B., Jin, H.S., Gottlieb, U., 2020. Customer service chatbots: anthropomorphism
Res. 161, 113838. and adoption. J. Bus. Res. 115, 14–24.
Marsh, S., Dibben, M.R., 2003. The role of trust in information science and technology. Shumanov, M., Johnson, L., 2021. Making conversations with chatbots more
Annu. Rev. Inform. Sci. Technol. 37, 465–498. personalized. Comput. Hum. Behav. 117, 106627.
Mayer, R.C., Davis, J.H., Schoorman, F.D., 1995. An integrative model of organizational Söderlund, M., 2006. Measuring customer loyalty with multi-item scales: a case for
trust. Acad. Manag. Rev. 20 (3), 709–734. caution. Int. J. Serv. Ind. Manag. 17 (1), 76–98.
McKnight, D.H., Choudhury, V., Kacmar, C., 2002. Developing and validating trust Ullman, D., Malle, B.F., 2017. Human-robot trust: just a button press away. In:
measures for e-commerce: an integrative typology. Inform. Syst. Res. 13 (3), Proceedings of the Companion of the 2017 ACM/IEEE International Conference on
334–359. Human-Robot Interaction - HRI ’17, pp. 309–310.
McKnight, D.H., Cummings, L.L., Chervany, N.L., 1998. Initial trust formation in new van Wezel, M., Croes, E.A., Antheunis, M.L., 2021. "I’m here for you": can social chatbots
organizational relationships. Acad. Manag. Rev. 23 (3), 473–490. truly support their users? A literature review. In: International Workshop on Chatbot
McKnight, D.H., Carter, M., Thatcher, J.B., Clay, P.F., 2011. Trust in a specific Research and Design. Springer, Cham, pp. 96–113.
technology: an investigation of its components and measures. ACM Transactions on Wang, X., Lin, X., Shao, B., 2022. How does artificial intelligence create business agility?
Management Information Systems (TMIS) 2 (2), 1–25. Evidence from chatbots. Int. J. Inf. Manag. 66, 102535.
McLean, G., Osei-Frimpong, K., Barhorst, J., 2021. Alexa, do voice assistants influence Wiggers, K., 2023. IBM Intros a Slew of New AI Services, Including Generative Models.
consumer brand engagement?–Examining the role of AI powered voice assistants in TechCrunch. Retrieved from. https://techcrunch.com/2023/05/09/ibm-intros-a-sle
influencing consumer brand engagement. J. Bus. Res. 124, 312–328. w-of-new-ai-services-including-generative-models/.
Meuter, M.L., McCabe, D.B., Curran, J.M., 2013. Electronic word-of-mouth versus Xu, K., Chen, X., Huang, L., 2022. Deep mind in social responses to technologies: a new
interpersonal word-of-mouth: are all forms of word-of-mouth equally influential? approach to explaining the Computers are Social Actors phenomena. Comput. Hum.
Serv. Market. Q. 34 (3), 240–256. Behav. 134, 107321.
Morgan, R.M., Hunt, S.D., 1994. The commitment-trust theory of relationship marketing. Zaichkowsky, J.L., 1985. Measuring the involvement construct. J. Consum. Res. 12 (3),
J. Market. 58 (3), 20–38. 341–352.
Nass, C., Moon, Y., 2000. Machines and mindlessness: social responses to computers. Zhang, J., 2010. The sound of silence: observational learning in the US kidney market.
J. Soc. Issues 56 (1), 81–103. Market. Sci. 29 (2), 315–335.
Nisbett, R.E., Wilson, T.D., 1977. The halo effect: evidence for unconscious alteration of
judgments. J. Pers. Soc. Psychol. 35 (4), 250–256.
14