You are on page 1of 14

Electronic Commerce Research and Applications 58 (2023) 101250

Contents lists available at ScienceDirect

Electronic Commerce Research and Applications


journal homepage: www.elsevier.com/locate/elerap

Categorizing affective response of customer with novel explainable


clustering algorithm: The case study of Amazon reviews
Wonjoon Kim a, 1, Keonwoo Nam b, c, 1, Youngdoo Son b, c, *
a
Division of Future Convergence (HCI Science Major), Dongduk Women’s University 60 Hwarang-ro 13-gil, Seongbuk-gu, Seoul 02748, South Korea
b
Department of Industrial and Systems Engineering, Dongguk University - Seoul, 30 Pildong-Ro 1-Gil, Jung-Gu, Seoul 04620, South Korea
c
Data Science Laboratory (DSLAB), Dongguk University - Seoul, 30 Pildong-Ro 1-Gil, Jung-Gu, Seoul 04620, South Korea

A R T I C L E I N F O A B S T R A C T

Keywords: Electronic word of mouth (e-WOM) influences consumer decision-making. Since consumers’ affective experi­
Attention mechanism ences for products are vast, research is needed to understand and categorize them accurately. In this paper, we
Explainable artificial intelligence developed a deep learning-based clustering algorithm for categorizing consumer sentiment in product reviews
Sentiment analysis
and explored the applicability of this algorithm. A Deep Attentive Self-Organizing Map (DASOM) was created by
Electronic word of mouth
noting individualized sentimental characteristics of each review and interpreting why each review was included
in a particular cluster. As a result of analyzing 4941 reviews of Amazon, one of online commerce platforms, it
was confirmed that sentiment classification through DASOM could be effectively used to categorize implicit
affective experiences of consumers. DASOM was effective in identifying the relationship between multi-
dimensional affective elements that were difficult to derive from TF-IDF. Using the proposed methodology, it
is expected to provide practical information for companies that design products considering consumer affection.

1. Introduction Turetken, 2020; Liu & Chen, 2022). The most representative problem is
the bias of a star rating. The inconsistency between a star rating and a
Word of mouth (WOM) refers to personal communication between review is one of frequently mentioned issues. For example, one may
individuals concerning their perception of goods and services (Ye et al., encounter a high star rating with a review having negative contents. One
2009). WOM has traditionally been used by consumers to learn about may also encounter a low star rating for a review with favorable con­
products and services and evaluate their attributes prior to purchase tents. In addition, it is difficult to know the underlying criterion for the
(Shaheen et al., 2019; Mishra et al., 2021). Since the emergence of e- evaluation or various opinions made by customers about the product
business, research has consistently shown that online or electronic word- based on the star rating alone. Finally, while e-commerce platforms such
of-mouth (e-WOM) plays an even more important role than traditional as Amazon and eBay allow users to rate a product they purchase, plat­
WOM in shaping consumer attitudes and behaviors (Litvin et al., 2008; forms such as social media do not provide a rating service.
Cheung & Thadani, 2012; Zhao et al., 2020). In addition, since e-WOM is To address these problems, many attempts have been made to un­
provided in various forms on e-commerce, social media, blogging sites, derstand the implicit needs and actual feelings of customers in text re­
and video platforms such as YouTube, a vast amount of content can views (Heinonen & Medberg, 2018; Puschmann & Powell, 2018; Jeong
spread faster than WOM. et al., 2019; Valdivia et al., 2019). Basically, the method of analyzing
Since e-WOM, which is propagated through unspecified masses for a sentiment in e-WOM is a lexicon-based approach (Lee et al., 2021),
specific product, contains general emotional information related to the which builds the meaning of words in the document as a dictionary and
product, it can be used to recognize consumers’ needs (Gregoriades & scores those words based on the constructed dictionary to judge senti­
Pampaka, 2020). The representative form of e-WOM can be divided into ment. Lexicon-based sentiment analysis has a disadvantage in that it
reviews written in verbal expression and ratings for products/services. requires a lot of effort to build a sentiment dictionary because the per­
Although a star rating makes it easy and intuitive to determine the formance is highly dependent on lexical resources such as sentiment
reputation or quality of a product, it has some problems (Al-Natour and dictionaries. To overcome this shortcoming, studies using machine

* Corresponding author.
E-mail addresses: wjkim@dongduk.ac.kr (W. Kim), namkw@dgu.ac.kr (K. Nam), youngdoo@dongguk.edu (Y. Son).
1
These authors equally contributed to this work.

https://doi.org/10.1016/j.elerap.2023.101250
Received 3 May 2022; Received in revised form 30 January 2023; Accepted 28 February 2023
Available online 5 March 2023
1567-4223/© 2023 Elsevier B.V. All rights reserved.
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

learning techniques such as artificial neural networks (ANN), support • A novel interpretable deep learning-based clustering algorithm, Deep
vector machine (SVM), Naive Bayes, relevance vector machine, and so Attentive Self-Organizing Map (DASOM), is proposed by adopting
on have been attempted (Son & Lee, 2016). Recently, hybrid-based the feed-forward attention to SOM.
research combining these two methods to improve classification per­ • Attention mechanism makes DASOM able to interpret the results of
formance has been actively conducted (Gupta and Joshi, 2020; Yang sentiment categorization in user reviews as well as to find the rep­
et al., 2020a; Yang et al., 2020b). resentations suitable for clustering.
Sentiments in most previous studies are categorized by defining af­ • When applying to the online reviews of recliners, DASOM can derive
fective factors such as positive/negative in advance. In such case, human affective experiences such as ’soft’ and ’large’, which were not
intervention is required for labeling of sentiments. In addition, there is a revealed in the existing clustering method SOM so that it can be used
limitation in clustering all documents into limited affective elements for improving affective quality and product development by inter­
(Ma et al., 2017). Since users express characteristics of the same product preting the relationship between representative factors for products.
in various ways, a vast and complex lexicon can appear in e-WOM
(Pessutto et al., 2020). Therefore, clustering is more appropriate than The remainder of this paper is organized as follows. Section 2 pro­
classifying a specific affective experience into several categories. vides a theoretical background of this study and concisely reviews pre­
Document clustering is a text mining technique that aims to recog­ vious studies on sentiment analysis, deep clustering, and explainable
nize patterns among documents and obtain information by dividing artificial intelligence (XAI) to describe gaps in our study. Section 3 ex­
them into groups based on contents. It has been actively used for plains the proposed model and its components. Sections 4 and 5 then
sentiment analysis of text data (Li & Liu, 2012; Wang et al., 2018; Zhang describe the methodology used in the study and show results derived
et al., 2021). Document clustering in sentiment analysis aims to derive from the model. Finally, discussion of the results, limitations, and future
meaningful sentiment information from clustering results after applying research of this study with concluding remarks are provided.
an algorithm such as k-means & hierarchical clustering, self-organizing
map (SOM), and so on to the document data vector (Jain & Nandanwar, 2. Related work
2014). Among them, SOM, which expresses high-dimensional data in a
low-dimensional grid through dimensionality reduction and clustering 2.1. Sentiment analysis using document clustering
to understand the patterns implied by the data, has been widely used in
recent years (Son & Kim, 2023). However, sentiment analysis using Affective experiences related to specific products and services are
these document clustering methods has shown poor performance multidimensionally embedded in text data such as reviews, or social
because it is usually performed in a fixed input space of features based networking services (SNS) mainly written by users (Kim et al., 2019;
on word counting. Choo and Kim, 2023). Document clustering generally uses
Therefore, researchers have proposed deep clustering methods to high-dimensional text data as input data after vectorization or embed­
improve performance (Kim et al., 2020; Bai et al., 2021). However, since ding to form a meaningful cluster. Since customers express character­
these methods clusters embedded vectors of a document, focusing only istics of the same product in various ways, vast and complex affective
on clustering results, they cannot explain the clustering results or major lexicons are bound to appear redundantly. Various research studies are
sentiment factors existing in individual documents. To avoid the prob­ actively conducted in this field to acquire affective information of cus­
lem in which the reliability of consumer sentiment information is low­ tomers through document clustering (Ma et al., 2017; Wu et al., 2020;
ered according to the results of clustering, it is necessary to explain the Abbasi-Moud et al., 2021).
process or reason for the clustering of documents. In general, sentiment analysis using document clustering is focused
Therefore, there is a need to develop an interpretable algorithm on classifying polarity, aspect, and sentiment score by performing
model to explain how each document responds sensitively to which feature extraction of documents using natural language processing
emotional factors in the document clustering process form a cluster (Wu (NLP) technology and calculating similarity between documents. Ran­
et al., 2020). An attention mechanism is utilized for interpretable domized clustering cuckoo search (RCCS) method that classifies docu­
learning algorithms (Bahdanau et al., 2014; Lee et al., 2023). This ments into three clusters (positive, neutral, negative) through
concept was proposed for the first time in machine translation. It Levenshtein distance and Cosine similarity (Ahmed et al., 2020), clus­
quantifies the degree to which a word is translated into new language tering based on sentiment score through similarity graph (Abbasi-Moud
attention to words in an existing language. Learning is performed while et al., 2021), polarity sentiment analysis using TF-IDF and k-means,
the output sequence includes the context of the input sequence. Due to sentiment analysis using unigram, TF-IDF based hierarchical clustering,
these characteristics, it has been applied to various deep learning algo­ and sentiment aspect cluster extraction using multilingual word em­
rithms such as machine translation, showing performance improvement beddings are representative studies.
(Lin et al., 2017; Kiela et al., 2018). There is a problem when applying existing methods focusing on
The primary goal of this study was to propose and verify an inter­ classifying predefined or limited types of affective elements in a recent
pretable clustering method for affective experiences of online word of situation where multi-dimensional and interrelated information is
mouth using the feedforward attention mechanism. In this study, Deep generated and expanded from media of various platforms. To solve these
Attentive Self-Organizing Map (DASOM), an interpretable clustering problems, various research studies have been performed to classify
algorithm, was proposed to solve the black box issue of the existing deep multidimensional emotional information in documents using deep
learning-based clustering method. It combines the feed-forward atten­ learning methods. Attempts have been made to propose an explainable
tion mechanism with SOM to provide representations suitable for clus­ model to solve the black box issue, one of the most representative issues
tering and interpretability. The denoising autoencoder is also employed for deep learning (Pessutto et al., 2020).
to find more robust representations. For verification, we employed Ito et al. (2020) have analyzed the prediction process of sentiment
product reviews from Amazon, one of the online commerce platforms. polarity score through Contextual Sentiment Neural Network (CSNN)
An experiment confirmed that the proposed algorithm could cluster using initialization propagation. Nguyen et al. (2021) have investigated
affective experiences for specific products. Furthermore, DASOM was the importance of variables in predicting LSTM and ensemble models for
also able to grasp the influence of words in clustering affective experi­ binary polarity classification using SHAP values. However, when per­
ences of product reviews and derive affective elements related to forming sentiment analysis using document clustering, no study has
product design features based on context vector values of affective vo­ directly interpreted the affective vocabulary that affects the formation
cabularies for each cluster. process of each cluster or suggested why the affective vocabulary is
The main contributions of this paper can be summarized as follows: included in the corresponding cluster.

2
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

2.2. Deep clustering of deep neural networks, have been reported.


Many studies are being conducted to interpret predictive models
When existing clustering method is applied to multidimensional through various attention mechanisms. Yang et al. (2016a), Yang et al.
data, the similarity between data is likely to become meaningless, (2016b) have successfully analyzed a document classification model by
making the clustering ineffective (Forest et al., 2019). To solve this extracting relatively active sentences and words from the document
problem, a deep clustering method has emerged. Deep clustering refers classification task through self-attention stacked in two layers. Baan
to a clustering method using deep learning techniques. It generally uses et al. (2019) have conducted a study showing that multi-head attention
a joint learning method of cluster learning and representation learning can partially interpret the process of document summarization. Tian
of input data (Hershey et al., 2016). In deep clustering, the represen­ et al. (2020) have proposed a method to automatically analyze the
tation of input data is learned in a form that is easy for clustering by Chinese word separation and part-of-speech (POS) tagging process
combining the reconstruction process of the autoencoder in the first half through two feed-forward attention models.
of the clustering algorithm (Min et al., 2018). Deep clustering has
evolved mainly into three categories of architecture since its first 3. Proposed clustering model
description by Hershey et al. (2016) to learned embedding values of a
spectral clustering process by optimizing the affinity matrix between This section describes our proposed model, Deep Attention Clus­
input data. The first category is an autoencoder-based method that is tering Self-Organizing Map (DASOM), including SOM, feedforward
most used as a deep clustering model (Xie et al., 2016; Chazan et al., attention mechanism, and denoising autoencoder as preliminaries.
2019; Diallo et al., 2021; Cai et al., 2022). This method can reduce the
dimension of input data using an encoder. The nonlinear relationship 3.1. Self-Organizing map (SOM)
between variables is then learned through a reconstruction process to
obtain a feasible feature space for clustering. The second category is a SOM is a type of clustering method that forms high-dimensional data
feedforward neural network-based method (Yang et al., 2016a; Yang in the form of a map by mapping and sorting them into low-dimensional
et al., 2016b; Caron et al., 2018; Wang et al., 2021). This method is nodes (See Fig. 1). Each node has its own node embedding vector (Aij )
characterized by performing representation learning of input data more and an index value (i,j) on the map, and each input vector is mapped to
robustly and quickly using a deep neural network that has pre-trained the best matching unit (BMU) having the most similar node embedding
classification or clustering tasks for large-scale datasets (image, text, vector. Finally, nodes form some clusters on a low-dimensional map
etc.). Due to this characteristic, models in this category are dominated through the entire node update. The map has the characteristic of pre­
by convolutional neural network (CNN) based architectures that can serving the positional relationship of the input variable as it is and re­
learn robust representations for large datasets. The third category is a flects the semantic relationship between the each of input vectors’
deep generative model-based method (Jiang et al., 2016; Lim et al., internal representations (Kohonen, 1990). SOM proceeds through the
2020; Yang et al., 2020a; Yang et al., 2020b). This method performs following steps:
clustering tasks and generates new sample data from obtained clusters.
For data generation, it utilizes generative neural networks such as Step 1. Entire node initialization: Initialize (i, j),.Aij
generative adversarial networks (GAN) (Goodfellow et al., 2014) and Step 2. BMU mapping and update: Calculate the node embedding
variational autoencoder (VAE) (Kingma & Welling, 2013). vector closest to the first input vector (X1 ) and update the corre­
sponding node (BMU) by the difference.
2.3. Explainable artificial intelligence based on attention mechanism
BMU = argminij ‖X1 − Aij ‖2 (1)
The attention mechanism (Bahdanau et al., 2014) refers to a method
ABMU ←ABMU + ‖X1 − ABMU ‖ (2)
that uses a vector to reflect the degree to which a specific output vector
pays attention to all input vectors as a weight. This mechanism was first
proposed to solve the problem that a fixed-length context vector could Step 3. Remaining node update: Update by reflecting a neighborhood
not contain sufficient information as the length of the sentence input to function (N ) together for preserving positional relationships in raw
RNNs (Recurrent Neural Networks)-based encoder-decoder model input space.
increased in the machine translation task. This study solved the problem ⃦ ⃦
by using the attention weight, which quantified how a specific word in Aij\BMU ←Aij\BMU + N ⃦X1 − Aij\BMU ⃦ (3)
the output language paid attention to the input word in context as a
context vector. When the attention mechanism is applied to a deep Step 4. Repeat step 2–4 until the last input vector (Xn ).
neural network model, it is known that interpretation of the working
principle inside the model is possible because the model can calculate SOM has been used for review text analysis because of its visuali­
the degree of attention to each token when performing a specific task zation ability of nodes where the text-embedded vectors are mapped, e.
(Chaudhari et al., 2021). g., the vectors generated from the TF-IDF algorithm. However, because
As the attention mechanism shows good performance in NLP, SOM uses the embedded representation vectors without any processing,
attention mechanism is widely used in various fields that utilize deep it may have difficulty finding good clustering for the texts with highly
neural networks, such as document classification (Yang et al., 2016a; different word distributions. In other words, it may generate some
Yang et al., 2016b; Lin et al., 2017; Kiela et al., 2018), speech recogni­ clusters with few instances or outliers, which induces meaningless
tion (Chorowski et al., 2015; Chan et al., 2016), image captioning (Xu clustering. These phenomena can become more severe in text analysis
et al., 2015), and recommendation systems (Ying et al., 2018). The with highly sparse embeddings.
attention mechanism has been modified into several forms according to Therefore, the DASOM proposed in this study tried to find the
the characteristics of the application target. Representatively, self- attention vector for each instance to obtain more suitable representa­
attention (Yang et al., 2016a; Yang et al., 2016b), which calculates the tions for clustering. In addition, these attention values are provided for
degree of attention and relationship between word tokens in sentences, each element in a representation vector so that it can be easily
multi-head attention, which connects self-attention in parallel (Vaswani interpreted.
et al., 2017), and feed-forward attention (Raffel & Ellis, 2015), which
continuously calculates weight values assigned to each input variable
through a separate feed-forward neural network in the learning process

3
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

Fig. 1. Principle of clustering of Self-Organizing Map.

3.2. Feedforward attention mechanism It is a variant of an autoencoder, taking an input corrupted with noise
and reconstructing a clean one. It is known that a denoising autoencoder
Fig. 2 shows a feedforward attention mechanism with a feedforward can learn a more robust representation of an input instance (Vincent
network, fθ, which takes an n-dimensional vector × as an input and et al., 2010; Tschannen et al., 2018; Shen et al., 2020). In particular,
generates an attention vector α representing each element’s importance when using real-world data, there are many cases where the data is noisy
weight in x. After α and × are combined, they are then transmitted to or unrefined, so a denoising autoencoder is used in various fields, such as
another deep neural network, gϕ, which conducts the target task. Thus, locomotion activities, image classification/segmentation, signal detec­
the feed-forward attention mechanism has a structure that a feed- tion, and recommendation system (Gu et al., 2018; Roy et al., 2018; Yu
forward neural network for finding attention values, fθ, is added to the et al., 2018; Chiang et al., 2019; Zhang et al. 2019).
first part of a network for the target task. Fig. 3 shows a brief structure of a denoising autoencoder. An
In details, the feedforward attention mechanism works as follows. instance × is corrupted with noise. The corrupted input ×̃becomes an
When an input × is given, the attention vector α is calculated as a input of an autoencoder which is trained to minimize the difference
feedforward network fθ(x). The sum of the elements in α is normalized between a clean instance × and the reconstructed input x’ so that it
by adding a soft-max layer at the end of fθ. A context vector c is calcu­ reconstructs the clean instance from the corrupted one. A typical choice
lated by conducting elementwise multiplication of α and x. It is then of noise is Gaussian or masking noise. As mentioned above, the
used as an input of the network for a target task, gϕ. Both networks are denoising autoencoder can learn a robust representation, z, due to
trained simultaneously so that elements in α can represent how much the noise’s enlarged manifold of inputs. Thus, we adopted the denoising
model focuses on each variable to perform a target task. autoencoder to find a robust representation of high-dimensional review
In this study, we focused on two features of the feed-forward atten­ data. Other deep clustering algorithms (Xie et al., 2016; Guo et al., 2018;
tion mechanism. Each input variable has its attention weight, and each Forest et al., 2019) usually employ a conventional autoencoder.
instance has its attention vector. Therefore, the feedforward attention
mechanism can provide how each variable in each sample affects the 3.4. DASOM: Deep Attentive Self-Organizing map
trained model. We proposed an interpretable clustering algorithm based
on this property by applying the feedforward attention mechanism to This study proposed a novel interpretable deep clustering algorithm
deep clustering. TF-IDF vectors of text documents were used as input called Deep Attentive Self-Organizing Map (DASOM) by combining a
instances so that the effect of each word in the document on clustering feedforward attention mechanism and a deep self-organizing map to
results could be presented. analyze customer sentiments from review texts. As shown in Fig. 4,

3.3. Denoising autoencoder

A denoising autoencoder was first proposed by Vincent et al. (2008).

Fig. 2. Illustrative of feed-forward attention mechanism. Fig. 3. General structure of denoising autoencoder.

4
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

DASOM consists of two parts: (1) a denoising autoencoder with a feed­ nodes on the map constructed by SOM as in Forest et al. (2019), where
forward attention mechanism and (2) a self-organizing map (SOM) of d(⋅, ⋅) was the distance between two node embeddings and T was the
the context vector, the input vector of which its attention score weights temperature parameter for controlling the bandwidth of the Gaussian
each element. First, the proposed method trains a denoising autoencoder function. The clustering loss with SOM can be written as follows:
after adding Gaussian noises to input instances. The penultimate layer of ∑I− 1 ∑J− 1 ⃒⃒ ⃒⃒2
the encoder part, a preceding layer of the innermost layer, was designed Lclu = i=0 j=0
N T (d(χ (z), ij) )⃒⃒z − mij ⃒⃒2 (5)
to have the same dimension of input to play the role of an attention
With the loss function in Eq. (5), embeddings of BMU and nearby
vector. A context vector was then calculated as the element-wise product
nodes are rapidly trained to be closer to the input, whilst nodes distant
of the original input vector and the corresponding attention vector. It
from an input are less affected. It is also noteworthy that the clustering
was used as an input of a self-organizing map that formed clusters of ∑ ∑ ⃒⃒ ⃒⃒2
those context vectors. The denoising autoencoder and the self- loss in Eq. (5) is converged to a k-means loss, I− 1 J− 1 ⃒⃒z − mχ(z) ⃒⃒ , if
i=0 j=0 2
organizing map were jointly trained. Thus, high weights were auto­ the temperature decreases to zero during training.
matically given to features of each input, highly affecting the clustering Finally, the total loss function is defined a weighted sum of two loss
result. The clustering result of each input then became explainable with functions, Lrec and Lclu , as follows:
attention scores of features.
Ltot = Lclu + γLrec (6)
Detailed procedures are as follows. An input vector, x ∈ Rd , is given
as a TF-IDF vector composed of d sentimental words. With a corrupted where γ is a weight parameter that controls the trade-off between two
input, ̃x = Gaussian Noise(x), and a reconstructed input, x = gϕ (fθ (̃ losses. The proposed DASOM model can be trained with the total loss

x)),
where fθ and gϕ denote the encoder and the decoder, respectively, the function in Eq. (6) by an end-to-end learning scheme. The overall al­
reconstruction loss of a denoising autoencoder can be defined as follows: gorithm of the proposed method is presented in [Algorithm 1].

(1)

Lrec = ||x − x ||22
Algorithm 1. Deep Attentive Self-Organizing Map (DASOM)
In addition, from the penultimate layer of the encoder part of the Given a dataset D = {x1 , x2 , ⋯, xN },
an encoder fθ and a decoder gϕ ,
denoising autoencoder, the attention vector α ∈ Rd and the context a network from input layer to penultimate layer of the encoder fθ\e ,
vector z ∈ Rd are calculated as follows: variance of Gaussian noise σ2 ,
the number of rows and columns of two-dimensional SOM (I, J),
α = softmax(fθ\e (̃x)) (2) the temperature of Gaussian neighbouring function T,
and the weight between two loss functions γ
z=α⊙x (3)
1: Repeat
2: for all xn ∈ D
where fθ\e (⋅) denotes the network from the input layer to the penultimate 3: xn = N (xn , σ2 I)
Sample a corrupted input ̃
( ( ))
and ⊙ denotes the elementwise multiplication of two vectors. 4:
Find a reconstruction loss Lrec = ||x − x ||22 where xn = gϕ fθ ̃
′ ′
xn
These context vectors are clustered by two-dimensional SOM with I ( )
5: Calculate an attention vector αn = softmax fθ\e (xn ) and a context vector zn =
rows and J columns. First, the best matching unit (BMU) of an input z,
αn ⊙ xn
χ (z), is defined as follows: 6: Find BMU of the context vector χ (zn ) = argminij ‖zn − mij ‖2
⃒⃒ ⃒⃒2 7: ∑ ∑ 1 T ⃒⃒ ⃒⃒2
χ (z) = argmini,j ⃒⃒z − mij ⃒⃒2 (4) Calculate a clustering loss Lclu = I−i=01 J−
j=0 N (d(χ (zn ), ij) ) zn − mij 2
⃒⃒ ⃒⃒
{ }
8: Update all parameters θ, ϕ, mij 0≤i≤I− 1,0≤j≤J− 1 with respect to Ltot = Lclu +
where mij is the embedding of ij-th node of SOM, 0 ≤ i ≤ I − 1 and γLrec
9: endfor
0 ≤ j ≤ J − 1. We employed Gaussian neighboring function,
10: Until converged.
d(⋅,⋅)2
N T (d(⋅, ⋅) ) = e− T2 , to define the topographic distance between two

Fig. 4. Overall structure of proposed algorithm (DASOM).

5
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

According to Besson (2016), training a SOM has the computational 5. Results


( )
complexity of O(nη dS +qS2 ) where n is the number of training sam­
ples, η is the maximal number of training epochs, d is the dimension of 5.1. Results of sentiment clustering by DASOM
an input vector, q is the dimension of an embedded vector for the SOM,
and S = I*J is the number of nodes in the SOM. Results of sentiment clustering for customer reviews using DASOM
Our method is an extension of SOM with some additional trainable are shown in Fig. 6. Numbers from 0 to 56 displayed inside the nodes
network parameters. Since the computational complexity of training a were index numbers sequentially assigned to each node. Corresponding
neural network with the number of weights, w, is O(nηw) (Zhu et al., nodes were BMU nodes to which one or more reviews were assigned.
2018), the complexity of training the proposed method can be summa­ The seven nodes not marked with a number were nodes to which no
( (
rized as O nη dS + qS2 + w
))
where w is the number of additional review data were assigned. In the color map, the inner color of each node
weight parameters in the denoising autoencoder. The computational reflected the average distance from neighboring nodes. The closer the
complexity of adding Gaussian noise can be omitted because it is pro­ distance between nodes, the lighter the color. On the other hand, the
portional to nηd. It is noteworthy that DASOM requires additional border color of each node was a criterion for categorizing clusters.
computation, which is proportional to w, compared to SOM to obtain the Agglomerative hierarchical clustering was used to set criteria for
attention scores. Hence, the additional computation can be controlled by generating clusters of nodes derived from DASOM. Clusters were
adjusting the size of the denoising autoencoder. recognized by considering each node embedding vector as a cluster and
merging clusters based on their similarity (Vesanto and Alhoniemi,
4. Methods 2000; Hormann & Fischer, 2019). In this study, clusters were identified
by dividing the dendrogram derived through agglomerative hierarchical
4.1. Experimental dataset and data Pre-Processing clustering using a static threshold. Clusters were then merged at a level
where the distance between clusters was less than 0.002 based on Ward
Customer review data registered by product consumers online Linkage. As can be seen in Fig. 7, the total number of clusters was 10,
contain information on the user’s emotional experience represented by with cluster 9 having the most reviews and cluster 6 having the fewest
product quality, price, design, and aesthetic satisfaction (Hennig-Thurau reviews.
et al., 2004). In this study, recliner product review data from Amazon.
com, one of the largest online e-commerce sites in the world, were 5.2. Comparison of TF-IDF and context vector for affective vocabulary
collected. The recliner is furniture that includes various emotional ele­
ments for users depending on the material and the product size (Kim For a total of ten clusters, results of extracting the five most affective
et al., 2019). vocabularies with the highest values of TF-IDF and context vector for
In this study, each review document was TF-IDF vectorized through each cluster are shown in Table 1 below. The ratio of the affective vo­
the data pre-processing process described by Kim et al. (2019). The data cabulary to total reviews of each cluster is indicated in parentheses.
pre-processing process was as follows. First, we obtained review data In Table 1, it could be seen that the value of attention TF-IDF was
and metadata on the Amazon website for the recliner through web generally higher than that of TF-IDF. This meant that the SOM increased
parsing. After tokenizing linguistic units such as words and documents, the influence of major affective vocabulary with high frequency for each
meaningless stop words in the documents were removed. In addition, cluster by the feed-forward attention mechanism when generating
POS tagging was performed to assign parts of speech, such as nouns and clusters while minimizing effects of other affective words on clustering.
adjectives, to each word constituting the document. The primary form
containing the word’s meaning was obtained by extracting the stem and 5.3. Results of sentiment clustering for sub-cluster
headword. Finally, a Document-Term Matrix was calculated based on
each document’s frequency of words with emotional meaning. A A dendrogram derived through agglomerative hierarchical clus­
component of the document-word matrix consists of the product of the tering did not provide a unique criterion for clustering. When clustering
term frequency (TF) in the document and the inverse document fre­ was performed based on data, some clusters might consist of multiple
quency (IDF) of the word frequency among documents. As a result, we sub-clusters. Partitioning at different levels for each branch in the
calculated a Document-Term Matrix consisting of 4,941 review data and dendrogram to perform rational partitioning on specific data has been
46 affective vocabularies for Amazon’s recliner products. proposed by Vesanto & Alhoniemi (2000).
The number of reviews belonging to cluster No. 9 out of ten clusters
4.2. Experimental design (procedure) derived based on the ward-linkage criteria used in this study was 1,602,
which was relatively large compared to other clusters. When many re­
In this study, hyperparameter values were experimentally set views are included in a cluster, it can be considered that the cluster
considering the number and dimensions of review data. First, the consists of core affective responses constituting the cluster with related
reconstruction loss function of the denoising autoencoder and the clus­ sub-level affective responses. Therefore, in this study, cluster 9 was
tering loss function of the self-organizing map were combined at a ratio additionally partitioned using a dynamic threshold. Results are shown in
of 1.5:1. The denoising autoencoder was pre-trained with 30 epochs Fig. 8 below. The value of dynamic threshold was 0.001.
after setting the standard deviation σ of Gaussian noise to be 0.5. The As a result of the partition using dynamic threshold, cluster 9 was
dimension of the self-organizing map was determined to be (8, 8), and T divided into three sub-groups as shown in Fig. 9. The first sub-cluster
representing the radius of the Gaussian neighbor function, was set to contained nodes No. 46/52/53/54. The second sub-cluster contained
exponentially decrease from a maximum of 10 to a minimum of 0.01. In nodes No. 48/55/56. The third sub-cluster contained nodes No. 34/35/
addition, learning the entire DASOM algorithm was carried out by 40/41/42/47.
setting the training epoch to 3,000 and the batch size to 256. The In these three sub-groups derived from cluster 9, five affective vo­
structure of the DASOM model used in the experiment is shown in Fig. 5 cabularies with the highest value of TF-IDF vector and context vector
below. were calculated. Results are shown in Table 2 below. It was confirmed
that the average score of context vector was generally higher than that of
TF-IDF in the subgroup, similar to results shown in Section 5.2. In the
case of clusters 9–1 and 9–3, it was confirmed that ‘comfort’ did not rank
in the top criteria for the value of TD-IDF vector, although it was
included in criterion for the value of context vector.

6
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

Fig. 5. The procedure of the DASOM model used in the experiment.

6. Discussion of the product (Fang et al., 2016). Therefore, accurately categorizing


core affective experience in customer reviews can be essential in product
6.1. Discussion of key findings design and advertising strategies. In addition, since the affective expe­
rience that occurs when a user uses a product is related to several
Analyzing and categorizing customer reviews is very important product attributes, it can be used to prioritize product development by
because a user’s emotional experience can help us understand the suc­ identifying their relationship. For example, in conjunction with the Kano
cess of the product in the market and the direction of future product model and the product life cycle, when a product corresponding to an
innovation and development. Customer review data include emotions early market is developed, if design elements or affective experiences
divided into positive, neutral, or negative. These emotions are config­ corresponding to attractive quality appear in customer reviews, it can be
ured in various forms, such as opposing, hierarchical, and compatible. expected to have a positive effect in the market (Kano, 1984).
The proposed clustering method utilizes a deep learning approach to Results of an investigation through a case study showed that various
categorize users’ emotions, find essential affective information about affective experiences for a product could be categorized more effectively
products, and understand and review their associations. with the proposed method than the conventional method. For example,
Forming a collective affective experience for a product displayed on the affective experience included in all clusters of data for a recliner was
an e-commerce platform can affect consumers’ evaluation and purchase ‘comfort,’ which could be seen as the most common and essential

7
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

Fig. 6. Results for analysis of hierarchical clustering.

Fig. 7. Results of sentiment clustering for customer reviews using DASOM.

affective experience with the recliner. In addition, the value of the demonstrative adjectives for recliner end-user reviews. As a result,
context vector compared to the value of TF-IDF in ‘comfort’ was found to ‘comfort’ was derived as the most appropriate affective adjective for a
increase for all clusters. This trend can be interpreted as a more signif­ recliner.
icant influence of ’comfort’ in all clusters by the forward attention
mechanism. Through this, it can be seen that when a customer using a
recliner expresses an experience about the product, emotional factors 6.2. Theoretical implications
related to subjective comfort are universally applied. These results are
consistent with those of previous studies. For example, Kim et al. (2018) Identifying the core value of a product by categorizing customers’
conducted a survey targeting experts and asked users using a Likert Scale emotions using the product can help develop a product based on a
widely used in the field of affective engineering. As a result, they derived positive affective experience of a customer. A user’s affective experience
that ‘comfort’ was the most critical emotional experience factor when can influence a company’s decision-making by identifying product
using a train seat. Kim et al. (2019) analyzed the frequency of characteristics affecting customer purchasing behavior. This study
attempted to categorize the affective experience of a product by using an

8
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

Table 1
Results of top 5 affective vocabularies based on the value of TF-IDF and context vector for each 10 clusters.
Affective Vocabulary Attractive Comfort Love Nice Perfect Soft

TF-IDF (Ratio) 0.32 0.18 (4.55%) – 0.19 (4.61%) 0.15 (3.73%) 1.08 (26.49%)
(7.85%)
Context value (Ratio) 0.0017 (8.15%) 0.0013 (6.24%) 0.0012 (5.59%) 0.0007 (3.49%) – 0.0088 (41.87%)

(a)

Affective Vocabulary Comfort Difficult Happy Love Nice Tight


TF-IDF (Ratio) 0.11 (3.53%) 0.62 (19.47%) 0.70 (22.19%) – 0.13 (4.24%) 0.10 (3.31%)
Context value (Ratio) 0.0007 (4.20%) 0.0048 (30.45%) 0.0048 (30.37%) 0.0007 (4.48%) 0.0007 (4.65%) –

(b)

Affective Vocabulary Best Comfort Disappoint Nice Poor Vibrate


TF-IDF (Ratio) 0.13 (6.72%) 0.10 (4.89%) 0.17 (8.53%) 0.52 (26.82%) – 0.23 (11.86%)
Context value (Ratio) 0.0009 (5.89%) 0.0011 (7.51%) – 0.0061 (40.14%) 0.0010 (6.26%) 0.0019 (12.33%)

(c)

Affective Vocabulary Cheap Comfort Hard High Long Nice Poor Inconvenient
TF-IDF (Ratio) 1.42 (43.03%) – – 0.12 (3.73%) 0.12 (3.71%) – 0.09 (2.70%) 0.11 (3.26%)
Context value (Ratio) 0.0086 (58.26%) 0.0005 (3.63%) 0.0009 (5.91%) – – 0.0003 (2.36%) 0.0004 (2.62%) –

(d)

Affective Vocabulary Comfort Love Nice Old Perfect Quiet


TF-IDF (Ratio) 0.10 (4.84%) 0.61 (30.71%) – 0.10 (5.00%) 0.16 (8.13%) 0.08 (4.07%)
Context value (Ratio) 0.0009 (5.97%) 0.0088 (56.04%) 0.0005 (3.38%) – 0.0009 (5.89%) 0.0007 (4.26%)

(e)

Affective Vocabulary Comfort Glad Love Perfect Satisfied Wonderful


TF-IDF (Ratio) 0.12 (3.70%) 0.77 (23.27%) – 0.12 (3.48%) 0.62 (18.62%) 0.43 (12.92%)
Context value (Ratio) 0.0008 (5.00%) 0.0057 (34.63%) 0.0006 (3.93%) – 0.0016 (9.98%) 0.0037 (22.47%)

(f)

Comfort Large Nice Perfect Pleased


Affective Vocabulary
TF-IDF (Ratio) 0.13 (4.05%) 0.86 (27.29%) 0.11 (3.53%) 0.15 (4.91%) 0.51 (16.15%)
Context value (Ratio) 0.0009 (6.08%) 0.0057 (39.24%) 0.0006 (3.88%) 0.0005 (3.51%) 0.0026 (18.22%)

(g)

Affective Vocabulary Comfort Durable Excellent Hard High Love Perfect


TF-IDF (Ratio) 0.12 (3.43%) 0.43 (12.12%) 1.16 (32.65%) – 0.12 (3.52%) – 0.11 (3.07%)
Context value (Ratio) 0.0006 (3.45%) 0.0020 (11.95%) 0.0086 (52.35%) 0.0004 (2.26%) – 0.0006 (3.70%) –

(h)

Affective Vocabulary Comfort High Long Perfect Sturdy


TF-IDF (Ratio) 0.19 (8.11%) 0.15 (6.44%) 0.12 (5.04%) 0.15 (6.59%) 0.13 (5.54%)
Context value (Ratio) 0.0022 (23.44%) 0.0007 (7.88%) 0.0006 (6.41%) 0.0011 (11.49%) 0.0007 (7.74%)

(i)

Affective Vocabulary Awesome Comfort Flat Hard Stiff


TF-IDF (Ratio) 0.30 (10.22%) 0.11 (3.69%) 0.21 (6.98%) 0.79 (26.87%) 0.15 (5.08%)
Context value (Ratio) 0.0018 (10.86%) 0.0008 (5.10%) 0.0013 (7.57%) 0.0066 (36.67%) 0.0013 (7.66%)
(j)

explainable artificial intelligence (XAI) technique and to identify the Therefore, in this study, we proposed DASOM, which applied an
relationship between users’ affective experience and the core charac­ attention mechanism to deep clustering for the first time. In DASOM,
teristics of the product. Machine or deep learning methods used in unlike DESOM, the role of each affective variable can be interpreted
sentiment analysis have the advantage of showing higher accuracy than through attention weight by using the size of the input vector of the SOM
traditional statistical approaches. Thus, they have been used in many in the same dimension as the input. As attention value can act as a
recent studies (Zhang et al., 2014; Do et al., 2019; Yadav & Vishwa­ weight for each input variable, it is possible to understand why input
karma, 2020). However, these techniques have the disadvantage that is variables belong to the corresponding cluster. This makes it easier to
expressing the causal relationship of categorization is difficult without interpret the cluster, and affective variables included therein.
providing adequate reliability. Since the neural network-based SOM To understand consumer purchasing behavior for a product, it is
model performs clustering using a pre-defined TF-IDF-based document important to observe significant differences in the complex emotions of
vector, there is a possibility that a suitable cluster cannot be found. In customers. For example, in the case of ‘hard’, it could mean that the
addition, the importance of each affective element may not be reflected. tactile property of a product is hard. It is also an affective term
Among these problems, the deep embedded self-organizing map expressing difficulty assembling or installing a product. Since many af­
(DESOM) concept was proposed to solve the cluster fit. In DESOM, SOM fective vocabularies have ambiguous expressions and meanings, as in
clustering was performed after embedding the input data in a space the case shown above, a DASOM that can suggest how affective terms
suitable for clustering by combining an autoencoder and SOM. However, are mentioned and the reason for categorization among affective vo­
since DESOM uses vectors embedded in the feature space, there is an cabularies in research related to sentiment analysis is needed.
issue that it is impossible to interpret each cluster individually (Forest
et al., 2019).

9
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

Fig. 8. Results of setting criteria for generating sub-clusters (ward linkage = 0.001).

Fig. 9. Result of dividing cluster 9 into sub-clusters.

6.3. Practical implications On the other hand, in the case of cluster 10, the tactile sensibility of
‘hard’ (36.67%), the opposite of ‘soft’, was categorized as a represen­
The results of this study suggest consistent outputs in terms of the tative affective experience of the cluster. Similar to ’stiff’ (7.66%),
applicability and suitability of DASOM for customer sentiment clus­ perceived as a somewhat negative expression regarding usability and
tering. First, it is possible to grasp the main affective elements of a tactile sensitivity, ’hard’ also revealed difficulties encountered when
specific product more clearly than TF-IDF. Using data sets applied in this installing or using the product (Kim et al., 2019; Kim, 2021). As such, it
study, it was confirmed that various affective factors could represent could be seen that reviews belonging to cluster 10 showed a complex
each cluster except for ‘comfort’ mentioned in the previous section. In affective experience. In addition, it was confirmed that even in group 2,
the case of cluster 1, the recliner was expressed as ’soft’, an affective a multi-dimensional emotion was implied when evaluating products,
vocabulary related to the tactile sense of the surface material. Looking at like in group 10. Many reviews pointed out the difficulty of assembling
tactile-related studies on materials and products, Chen et al. (2009) have or processing and using a specific function through the affective vo­
confirmed that ’slippery-sticky’ and ’hard-soft’ are the most necessary cabulary ‘difficult’. However, they expressed happiness about other
affective experiences of tactile preference for various types (37 types) of factors, such as color and cost-effectiveness. Therefore, the easier the
fabrics. ‘Soft’ and ‘comfort’ are affective words used to express the recliner’s assembly, processing, and function, the higher the evaluation
comfort level of a seat. ‘Attractive’ is an affective word used to describe a from users of Cluster 2.
unique design (Hapsari et al., 2017). In cluster 3, similarly to clusters 1 and 10, tactile sensitivity was

10
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

Table 2
Results of top 5 affective vocabularies based on the value of TF-IDF and context vector for sub-clusters in cluster 9.
Affective Vocabulary Comfort Heavy Long Old Quick Sturdy

TF-IDF (Ratio) – 0.19 0.32 (11.13%) 0.34 (11.79%) 0.18 (6.16%) 0.53 (18.18%)
(6.64%)
Context value (Ratio) 0.0012 (8.75%) 0.0010 (7.27%) 0.0024 (17.45%) 0.0022 (15.53%) – 0.0033 (23.43%)

(a)

Affective Vocabulary Comfort High Pretty Relax Simple Tight


TF-IDF (Ratio) 0.27 (22.78%) 0.09 0.09 (7.40%) 0.05 (4.05%) 0.07 (5.77%) –
(7.76%)
Context value (Ratio) 0.0037 (54.15%) 0.0005 (7.84%) 0.0004 (6.13%) – 0.0003 (4.87%) 0.0004 (5.58%)

(b)

Affective Vocabulary Beautiful Comfort High Perfect Pretty Inconvenient


TF-IDF (Ratio) 0.27 (7.57%) – 0.28 (7.80%) 0.38 (10.46%) 0.19 (5.21%) 0.14 (3.92%)
Context value (Ratio) 0.0013 (13.39%) 0.0006 (5.92%) 0.0013 (13.63%) 0.0031 (32.35%) – 0.0006 (6.68%)
(c)

important. The group evaluated the recliner as ‘nice’ (40.14%), an af­ composing group 3 mainly evaluated the massage function of the elec­
fective vocabulary with a positive meaning, and used the affective vo­ tric recliner, which was closely related to vibration. It could be under­
cabulary ‘vibrate’ (12.33%) as the basis. Unlike ‘soft’ or ‘hard’ for stood that most users expressed satisfaction through ‘nice’ regarding the
expressing the surface material, ‘vibrate’ is an affective vocabulary that vibration of this massage function.
describes the vibrotactile of an object (Asaga et al., 2013). Humans sense On the other hand, in cluster 4, ’cheap’ (58.26%) and ’hard’ (5.91%)
vibration, temperature, and material as emotional elements when pointed out the low quality of the product due to the size and incon­
interacting with an object (Jafar et al., 2018). In other words, reviews venience of use. In contrast, in cluster 7, ’large’ (39.24%) and ’pleased’

Fig. 10. Results of values of TF-IDF, Attention, and Context vector for affective vocabularies in each review.

11
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

(18.22%) confirmed that the product’s design elements were expressed proposed. DASOM forms a collaborative learning network with self-
positively. organizing maps and denoising autoencoders to prevent distortion of
In cluster 8, affective vocabularies such as ‘excellent’ (52.35%) and input data and perform correct clustering through the robustness of
‘durable’ (11.95%) mainly appeared. ’Durable’ is an affective vocabu­ representation learning. In addition, by applying the forward attention
lary when describing the user’s focus on safety when evaluating a chair mechanism inside the denoising autoencoder, the process and results of
(Rahman, 2012; Hapsaril et al., 2017). Based on reviews of cluster 8, it clustering can be interpreted by the adjustment effect of attention
was confirmed that the one-dimensional quality factor in the Kano weight.
model was critical in product evaluation. It is judged that users value the The effect of DASOM was verified through a case study on customer
safety of the recliner. Thus, the recliner product designer will need to reviews of the recliner on Amazon. Users’ sentiment toward the product
secure the level of quality expected by the user by improving its was analyzed. As a result of the analysis, it was confirmed that various
durability. factors related to the user’s experience, such as the design feature of the
Second, to verify the effectiveness of the DASOM algorithm in more product, usability, and subjective feeling, were complexly structured.
depth, this study analyzed individual reviews belonging to several Results demonstrate that DASOM helps derive each element into an
clusters. appropriate cluster. In particular, it was confirmed that the process of
Fig. 10 shows the TF-IDF, attention weight, and context vector score forming a corresponding cluster of individual reviews could be inter­
of individual reviews belonging to the cluster. Each figure results from preted based on the effect of the attention mechanism on affective vo­
reviews No. 15, 255, 1663, 18, and 394. Review No. 15, belonging to cabularies. DASOM confirmed that patterns of affective elements that
cluster 1, had high TF-IDF values for ’relax’ and ’solid’ but was set to were difficult to detect in TF-IDF commonly used in studies related to
have high z-score values for ’soft’ based on attention weight (Fig. 10 sentiment analysis through existing NLP could be more effectively
(a)). In other words, if clustering was carried out using the existing TF- clustered. Affective factors of users for each group derived from this
IDF, it was reasonable that review 15 was assigned to the cluster for study are expected to provide critical information on customizing
‘relax’ or ‘solid’ rather than ‘soft’. However, due to the effect of the essential for product design.
attention mechanism, the clustering algorithm contributed more to The limitations of this study are as follows. First, this study used
forming a cluster for ‘soft’ by paying more attention to ‘soft’. The impact Amazon data as a case study. Although Amazon is the largest commerce
of the attention mechanism could also be confirmed in (b) and (c) in platform in the world, it cannot represent all data. Therefore, it is
Fig. 10. necessary to conduct research using data from various platforms in the
Review number 255 was reversed from ‘awesome’ to ‘love’. Review future. In addition, the data set included English reviews only. Even for
number 1663 was assigned to clusters 5 and 2, respectively, from the same product, differences in affective experience may occur for
‘awesome’ to ‘happy’. What was noteworthy here was that existing TF- different languages and cultures. Thus, future work should consider the
IDF values of Review No. 255 and Review No. 1663 belonging to Cluster development of a sentiment classification method by cross-verifying
5 and Cluster 2 were very similar. Values for ‘awesome’ and ‘happy’ data from multiple cultures. Based on the results of this study, it is
were the same. However, the value of ‘love’, which existed only in re­ necessary to discuss the general use of the methodology for sentiment
view number 255, had a much lower influence. According to the TF-IDF clustering by classifying star ratings or positive/negative customer
standard, it is a reasonable clustering result to be grouped into the same reviews.
emotional clusters for ‘awesome’ or ‘happy’. However, DASOM assigned
review No. 1663 to the ‘happy’ cluster (cluster 2), whereas review No. CRediT authorship contribution statement
255 was assigned to the ‘love’ cluster (cluster 5). This overcame the
limitation of TF-IDF, which gave a low value to ‘love’ because it was a Wonjoon Kim: Conceptualization, Validation, Formal analysis,
frequent word in the entire review. It can be seen as an example of Investigation, Data curation, Writing – original draft, Writing – review &
confirming that DASOM can effectively analyze sentiment through more editing, Funding acquisition. Keonwoo Nam: Methodology, Validation,
detailed clustering. Writing – original draft. Youngdoo Son: Conceptualization, Writing –
Finally, (d) and (e) of Fig. 10 show three figures for each review No. review & editing, Funding acquisition.
18 and No. 394 included in the emotional cluster represented by ‘large’.
Each review had a very different TF-IDF vector. Review number 18 Declaration of Competing Interest
showed the highest number in the order of ‘disappoint’, ‘difficult’, and
‘large’. Review number 394 showed the highest number in the order of The authors declare that they have no known competing financial
‘pleased’, ‘large’, and ‘love’. According to the TF-IDF standard, it was interests or personal relationships that could have appeared to influence
reasonable for reviews No. 18 and No. 394 to belong to different clus­ the work reported in this paper.
ters, contrary to (b) and (c) of Fig. 10. However, DASOM assigned both
reviews to cluster 7, paying attention to ‘large’. The cluster contained Data availability
design sensibility related to the size of the recliner.
Results for the three sub-groups derived from cluster 9 confirmed Data will be made available on request.
that the first cluster was categorized into affective vocabularies such as
‘sturdy’ and ‘long’. In the second cluster, the recliner was recognized Acknowledgment
with affective vocabularies such as ‘comfort’ and ‘high’. In the last
cluster, affective vocabularies such as ‘high’ and ‘beautiful’ were mainly This research was supported by the National Research Foundation of
expressed along with ‘perfect’. ‘Perfect’ is an affective vocabulary Korea (NRF) grant funded by the Korea government (MSIT: Ministry of
frequently used when users of the recliner feel comfortable (Kim et al., Science and ICT) (No. 2020R1G1A1003384, RS-2023-00208412) and
2019). ‘Beautiful’ is an affective vocabulary that best expresses aesthetic also by Korea Institute of Planning and Evaluation for Technology in
excellence in product image evaluation. Therefore, it can be seen that Food, Agriculture, Forestry (IPET) through Smart Agri Products Flow
comfort and visual image are essential to attract group users to Storage Technology Development Program, funded by Ministry of
customers. Agriculture, Food and Rural Affairs (MAFRA) of Korea (No.322050-3).

7. Conclusion and future research direction

In this study, DASOM, an interpretable clustering algorithm, was

12
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

References Jafar, F.A., Rosli, N.A., Rashid, S.A., Ali, M.M., Shamsuddin, S., Kassim, A.M., 2018. In:
March). Towards the development of kansei haptic sensing technology for robot
application–Exploring human haptic emotion. Springer, Singapore, pp. 622–630.
Abbasi-Moud, Z., Vahdat-Nejad, H., & Sadri, J. (2021). Tourism recommendation system
Jain, Y., Nandanwar, A.K., 2014. A theoretical study of text document clustering. Int. J.
based on semantic clustering and sentiment analysis. Exp. Syst. Appl., 167, 114324.
Comput. Sci. Inf. Technol. 5 (2), 2246–2251.
Ahmed, S.H., Wassif, K.T., Nabil, E., 2020. Clustering Based Sentiment Analysis Using
Jeong, B., Yoon, J., Lee, J.M., 2019. Social media mining for product planning: A product
Randomized Clustering Cuckoo Search Algorithm. Int. J. Comput. Sci. Network
opportunity mining approach based on topic modeling and sentiment analysis. Int. J.
Secur. 20 (7), 159.
Inf. Manag. 48, 280–290.
Al-Natour, S., & Turetken, O. (2020). A comparative assessment of sentiment analysis
Jiang, Z., Zheng, Y., Tan, H., Tang, B., & Zhou, H. (2016). Variational deep embedding:
and star ratings for consumer reviews. Int. J. Informat. Manage., 54, 102132.
An unsupervised and generative approach to clustering. arXiv preprint arXiv:
Asaga, E., Takemura, K., Maeno, T., Ban, A., Toriumi, M., 2013. Tactile evaluation based
1611.05148.
on human tactile perception mechanism. Sens. Actuators, A 203, 69–75.
Kano, N. (1984). Attractive quality and must-be quality. Hinshitsu (Quality, J. Japan. Soc.
Baan, J., ter Hoeve, M., van der Wees, M., Schuth, A., & de Rijke, M. (2019).
Quality Control), 14, 39-48.
Understanding multi-head attention in abstractive summarization. arXiv preprint
Kiela, D., Wang, C., & Cho, K. (2018). Dynamic meta-embeddings for improved sentence
arXiv:1911.03898.
representations. arXiv preprint arXiv:1804.07983.
Bahdanau, D., Cho, K., & Bengio, Y. (2014). Neural machine translation by jointly
Kim, W., Lee, Y., Lee, J.H., Shin, G.W., Yun, M.H., 2018. A comparative study on designer
learning to align and translate. arXiv preprint arXiv:1409.0473.
and customer preference models of leather for vehicle. Int. J. Ind. Ergon. 65,
Bai, R., Huang, R., Chen, Y., Qin, Y., 2021. Deep multi-view document clustering with
110–121.
enhanced semantic embedding. Inf. Sci. 564, 273–287.
Kim, W., Ko, T., Rhiu, I., Yun, M.H., 2019. Mining affective experience for a kansei
Besson, L. (2016). Self-Organizing Maps and DSOM: From unsupervised clustering
design study on a recliner. Appl. Ergon. 74, 145–153.
algorithms to models of cortical plasticit. In Research project, Ecole Normale Supérieure
Kim, J., Yoon, J., Park, E., Choi, S., 2020. Patent document clustering with deep
de Cachan. (http://perso.crans.org/besson/publis/mva-2016/MVA_2015-16__Neuro-
embeddings. Scientometrics 123 (2), 563–577.
Sciences__Project__Lilian_Besson__Report.en.pdf).
Kim, W. (2021). A study on the subjective feeling affecting tactile satisfaction of leather
Cai, J., Wang, S., Xu, C., & Guo, W. (2022). Unsupervised deep clustering via contractive
in automobile: A structural equation modeling approach. Int. J. Indust. Ergonom., 84,
feature representation and focal loss. Patt. Recogn., 123, 108386.
103167.
Caron, M., Bojanowski, P., Joulin, A., & Douze, M. (2018). Deep clustering for
Kingma, D. P., & Welling, M. (2013). Auto-encoding variational bayes. arXiv preprint
unsupervised learning of visual features. In Proceedings of the European conference on
arXiv:1312.6114.
computer vision (ECCV) (pp. 132-149).
Kohonen, T., 1990. The self-organizing map. Proc. IEEE 78 (9), 1464–1480.
Chan, W., Jaitly, N., Le, Q., & Vinyals, O. (2016, March). Listen, attend and spell: A
Lee, S.W., Jiang, G., Kong, H.Y., Liu, C., 2021. A difference of multimedia consumer’s
neural network for large vocabulary conversational speech recognition. In 2016 IEEE
rating and review through sentiment analysis. Multimed. Tools Appl. 80 (26),
international conference on acoustics, speech and signal processing (ICASSP) (pp. 4960-
34625–34642.
4964). IEEE.
Lee, S., Park, H., Choi, C., Kim, W., Kim, K.K., Han, Y.-K., Kang, J., Kang, C.-J., Son, Y.,
Chaudhari, S., Mithal, V., Polatkan, G., Ramanath, R., 2021. An attentive survey of
2023. Multi-order graph attention network for water solubility prediction and
attention models. ACM Trans. Intellig. Syst. Technol. (TIST) 12 (5), 1–32.
interpretation. Scientific Reports 13, 957.
Chazan, S. E., Gannot, S., & Goldberger, J. (2019, October). Deep clustering based on a
Li, G., Liu, F., 2012. Application of a clustering method on sentiment analysis. J. Inf. Sci.
mixture of autoencoders. In 2019 IEEE 29th International Workshop on Machine
38 (2), 127–139.
Learning for Signal Processing (MLSP) (pp. 1-6). IEEE.
Lim, K.L., Jiang, X., Yi, C., 2020. Deep clustering with variational autoencoder. IEEE
Chen, X., Shao, F., Barnes, C.J., Childs, T., Henson, B., 2009. Exploring relationships
Signal Process Lett. 27, 231–235.
between touch perception and surface physical properties. Int. J. Des. 3 (2), 67–76.
Lin, Z., Feng, M., Santos, C. N. D., Yu, M., Xiang, B., Zhou, B., & Bengio, Y. (2017). A
Cheung, C.M., Thadani, D.R., 2012. The impact of electronic word-of-mouth
structured self-attentive sentence embedding. arXiv preprint arXiv:1703.03130.
communication: A literature analysis and integrative model. Decis. Support Syst. 54
Litvin, S.W., Goldsmith, R.E., Pan, B., 2008. Electronic word-of-mouth in hospitality and
(1), 461–470.
tourism management. Tour. Manag. 29 (3), 458–468.
Chiang, H.T., Hsieh, Y.Y., Fu, S.W., Hung, K.H., Tsao, Y., Chien, S.Y., 2019. Noise
Liu, X. X., & Chen, Z. Y. (2022). Service quality evaluation and service improvement
reduction in ECG signals using fully convolutional denoising autoencoders. IEEE
using online reviews: A framework combining deep learning with a hierarchical
Access 7, 60806–60813.
service quality model. Electron. Comm. Res. Appl., 54, 101174.
Chorowski, J. K., Bahdanau, D., Serdyuk, D., Cho, K., & Bengio, Y. (2015). Attention-
Ma, B., Yuan, H., Wu, Y., 2017. Exploring performance of clustering methods on
based models for speech recognition. Adv Neur. Informat. Process. Syst., 28.
document sentiment analysis. J. Inf. Sci. 43 (1), 54–74.
Choo, S., Kim, W., 2023. A study on the evaluation of tokenizer performance in natural
Min, E., Guo, X., Liu, Q., Zhang, G., Cui, J., Long, J., 2018. A survey of clustering with
language processing. Applied Artificial Intelligence 37 (1), 2175112.
deep learning: From the perspective of network architecture. IEEE Access 6,
Diallo, B., Hu, J., Li, T., Khan, G.A., Liang, X., Zhao, Y., 2021. Deep embedding clustering
39501–39514.
based on contractive autoencoder. Neurocomputing 433, 96–107.
Mishra, A., Shukla, A., Sharma, S.K., 2021. Psychological determinants of users’ adoption
Do, H.H., Prasad, P.W.C., Maag, A., Alsadoon, A., 2019. Deep learning for aspect-based
and word-of-mouth recommendations of smart voice assistants. Int. J. Inf. Manag.
sentiment analysis: a comparative review. Expert Syst. Appl. 118, 272–299.
102413.
Fang, J., Shao, Y., Wen, C., 2016. Transactional quality, relational quality, and consumer
Nguyen, A.n., Foerstel, S., Kittler, T., Kurzyukov, A., Schwinn, L., Zanca, D., Hipp, T.,
e-loyalty: Evidence from SEM and fsQCA. Int. J. Inf. Manag. 36 (6), 1205–1217.
Jun, S.D., Schrapp, M., Rothgang, E., Eskofier, B., 2021. System Design for a Data-
Forest, F., Lebbah, M., Azzag, H., & Lacaille, J. (2019). Deep embedded SOM: joint
Driven and Explainable Customer Sentiment Monitor Using IoT and Enterprise Data.
representation learning and self-organization. In European Symposium on Artificial
IEEE Access 9, 117140–117152.
Neural Networks, Computational Intelligence and Machine Learning (ESANN)..
Pessutto, L. R. C., Vargas, D. S., & Moreira, V. P. (2020). Multilingual aspect clustering
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., .. &
for sentiment analysis. Knowled.-Based Syst., 192, 105339.
Bengio, Y. (2014). Generative adversarial nets. Adv. Neural Informat. Process. Syst.,
Puschmann, C., & Powell, A. (2018). Turning words into consumer preferences: How
27.
sentiment analysis is framed in research and the news media. Social Media+
Gregoriades, A., & Pampaka, M. (2020). Electronic word of mouth analysis for new
Society, 4(3), 2056305118797724.
product positioning evaluation. Electron. Comm. Res. Appl., 42, 100986.
Raffel, C., & Ellis, D. P. (2015). Feed-forward networks with attention can solve some
Gu, F., Khoshelham, K., Valaee, S., Shang, J., Zhang, R., 2018. Locomotion activity
long-term memory problems. arXiv preprint arXiv:1512.08756.
recognition using stacked denoising autoencoders. IEEE Internet Things J. 5 (3),
Rahman, O., 2012. The influence of visual and tactile inputs on denim jeans evaluation.
2085–2093.
Int. J. Des. 6 (1), 11–25.
Gupta, I., Joshi, N., 2020. Enhanced twitter sentiment analysis using hybrid approach
Roy, S.S., Hossain, S.I., Akhand, M.A.H., Murase, K., 2018. A robust system for noisy
and by accounting local contextual semantic. J. Intell. Syst. 29 (1), 1611–1625.
image classification combining denoising autoencoder and convolutional neural
Hapsari, S.N., Sjafrizal, T., Anugraha, R.A., Nik Hisyamudin, M.N., Amir, K., Al Emran, I.,
network. Int. J. Adv. Comput. Sci. Appl. 9 (1), 224–235.
Mohd Rasidi, I., Mohd Faizal, M.B., Mohd Azlis Sani, M.J., Ahmad Mubarak, T.A.,
Shaheen, M., Zeba, F., Chatterjee, N., Krishnankutty, R., 2019. Engaging customers
Izzuddin, Z., Sofian, M., 2017. Designing Train Passenger Seat by Kansei Engineering
through credible and useful reviews: the role of online trust. YC 21 (2), 137–153.
in Indonesia. MATEC Web Conf. 135, 00017.
Shen, T., Mueller, J., Barzilay, R., & Jaakkola, T. (2020, November). Educating text
Heinonen, K., Medberg, G., 2018. Netnography as a tool for understanding customers:
autoencoders: Latent representation guidance via denoising. In International
implications for service research and practice. J. Serv. Mark. 32 (6), 657–679.
Conference on Machine Learning (pp. 8719-8729). PMLR.
Hennig-Thurau, T., Gwinner, K.P., Walsh, G., Gremler, D.D., 2004. Electronic word-of-
Son, Y., & Kim, W. (2023). Development of methodology for classification of user
mouth via consumer-opinion platforms: what motivates consumers to articulate
experience (UX) in online customer review. J. Retail. Consum. Serv., 71, 103210.
themselves on the internet? J. Interact. Mark. 18 (1), 38–52.
Tian, Y., Song, Y., Ao, X., Xia, F., Quan, X., Zhang, T., & Wang, Y. (2020, July). Joint
Hershey, J. R., Chen, Z., Le Roux, J., & Watanabe, S. (2016, March). Deep clustering:
Chinese word segmentation and part-of-speech tagging via two-way attentions of
Discriminative embeddings for segmentation and separation. In 2016 IEEE
auto-analyzed knowledge. In Proceedings of the 58th Annual Meeting of the Association
International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 31-
for Computational Linguistics (pp. 8286-8296).
35). IEEE.
Tschannen, M., Bachem, O., & Lucic, M. (2018). Recent advances in autoencoder-based
Hormann, R., & Fischer, E. (2019). Detecting Anomalies by using Self-Organizing Maps
representation learning. arXiv preprint arXiv:1812.05069.
in Industrial Environments. In ICISSP (pp. 336-344).
Son, Y., Lee, J., 2016. Active learning using transductive sparse Bayesian regression.
Ito, T., Tsubouchi, K., Sakaji, H., Yamashita, T., Izumi, K., 2020. Contextual sentiment
Information Sciences 374, 240–254.
neural network for document sentiment analysis. Data Sci. Eng. 5 (2), 180–192.

13
W. Kim et al. Electronic Commerce Research and Applications 58 (2023) 101250

Valdivia, A., Hrabova, E., Chaturvedi, I., Luzón, M.V., Troiano, L., Cambria, E., Yang, J., Parikh, D., & Batra, D. (2016). Joint unsupervised learning of deep
Herrera, F., 2019. Inconsistencies on TripAdvisor reviews: A unified index between representations and image clusters. In Proceedings of the IEEE conference on computer
users and Sentiment Analysis Methods. Neurocomputing 353, 3–16. vision and pattern recognition (pp. 5147-5156).
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Polosukhin, I., Yang, Z., Yang, D., Dyer, C., He, X., Smola, A., & Hovy, E. (2016, June). Hierarchical
2017. Attention is all you need. Adv. Neural Inf. Proces. Syst. 30. attention networks for document classification. In Proceedings of the 2016 conference
Vesanto, J., Alhoniemi, E., 2000. Clustering of the self-organizing map. IEEE Trans. of the North American chapter of the association for computational linguistics: human
Neural Netw. 11 (3), 586–600. language technologies (pp. 1480-1489).
Vincent, P., Larochelle, H., Bengio, Y., Manzagol, P.A., 2008. July). Extracting and Yang, L., Li, Y., Wang, J., Sherratt, R.S., 2020. Sentiment analysis for E-commerce
composing robust features with denoising autoencoders. In: In Proceedings of the 25th product reviews in Chinese based on sentiment lexicon and deep learning. IEEE
international conference on Machine learning, pp. 1096–1103. Access 8, 23522–23530.
Vincent, P., Larochelle, H., Lajoie, I., Bengio, Y., Manzagol, P.A., Bottou, L., 2010. Ye, Q., Law, R., Gu, B., 2009. The impact of online user reviews on hotel room sales. Int.
Stacked denoising autoencoders: Learning useful representations in a deep network J. Hosp. Manag. 28 (1), 180–182.
with a local denoising criterion. J. Mach. Learn. Res. 11 (12). Ying, H., Zhuang, F., Zhang, F., Liu, Y., Xu, G., Xie, X., .. & Wu, J. (2018, January).
Wang, Y., Lu, X., Tan, Y., 2018. Impact of product attributes on customer satisfaction: An Sequential recommender system based on hierarchical attention network. In IJCAI
analysis of online reviews for washing machines. Electron. Commer. Res. Appl. 29, International Joint Conference on Artificial Intelligence.
1–11. Yu, J., Huang, D., Wei, Z., 2018. Unsupervised image segmentation via Stacked
Wang, Z., Ni, Y., Jing, B., Wang, D., Zhang, H., Xing, E., 2021. DNB: A Joint Learning Denoising Auto-encoder and hierarchical patch indexing. Signal Process. 143,
Framework for Deep Bayesian Nonparametric Clustering. IEEE Trans. Neural 346–353.
Networks Learn. Syst. Zhang, H., Gan, W., & Jiang, B. (2014, September). Machine learning and lexicon based
Wu, P., Li, X., Shen, S., & He, D. (2020). Social media opinion summarization using methods for sentiment classification: A survey. In 2014 11th web information system
emotion cognition and convolutional neural networks. Int. J. Informat. Manage., 51, and application conference (pp. 262-265). IEEE.
101978. Zhang, J., Lu, X., & Liu, D. (2021). Deriving customer preferences for hotels based on
Xie, J., Girshick, R., & Farhadi, A. (2016, June). Unsupervised deep embedding for aspect-level sentiment analysis of online reviews. Electronic Commerce Research and
clustering analysis. In International conference on machine learning (pp. 478-487). Applications, 49, 101094.
PMLR. Zhang, C., Li, T., Ren, Z., Hu, Z., Ji, Y., 2019. Taxonomy-aware collaborative denoising
Xu, K., Ba, J., Kiros, R., Cho, K., Courville, A., Salakhudinov, R., .. & Bengio, Y. (2015, autoencoder for personalized recommendation. Appl. Intell. 49 (6), 2101–2118.
June). Show, attend and tell: Neural image caption generation with visual attention. Zhao, Y., Wang, L., Tang, H., & Zhang, Y. (2020). Electronic word-of-mouth and
In International conference on machine learning (pp. 2048-2057). PMLR. consumer purchase intentions in social e-commerce. Electron. Comm. Res. Appl., 41,
Yadav, A., Vishwakarma, D.K., 2020. Sentiment analysis using deep learning 100980.
architectures: a review. Artif. Intell. Rev. 53 (6), 4335–4385. Zhu, Q., Wang, S., Cheng, B., Sun, Q., Yang, F., Chang, R.N., 2018. Context-aware group
Yang, X., Deng, C., Wei, K., Yan, J., Liu, W., 2020. Adversarial learning for robust deep recommendation for point-of-interests. IEEE Access 6, 12129–12144.
clustering. Adv. Neural Inf. Proces. Syst. 33, 9098–9108.

14

You might also like