Professional Documents
Culture Documents
Open Access. © 2021 Liangfeng Qian, Shengli Deng, published by Sciendo. This work is licensed under the Creative Commons
Attribution-NonCommercial-NoDerivatives 3.0 License.
An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 373
Table 1
Comparison of Knowledge Aggregation Methods
Metadata Describe the property of data and realize the unified Simple and easy to use, Strong
integration of heterogeneous knowledge resources standardization, Weak semantics
Ontology-based Formal description of concept system to improve the Strong standardization, Formalized and
machine-readable and understandable data conceptualized, Semantic relevance
Associated data Naming network objects with uniform resource identifier Reveal the semantic meaning and relationship
and data publishing and resource association through HTTP of information to a certain extent
protocol
Social tag and cluster Simple and easy to use, strong freedom, business Poor standardization, Loose structure and
analysis collaboration fuzzy semantics
Knowmetrics Large amount of data processing, multi-dimensional and Weak semantics, Relying on auxiliary tools and
visualization methods
combines the relevance computed from several samples learning methods. Representation learning is “learning
as learned by a neural network and generates scores for representations of the data that make it easier to extract
each input feature. For poorly trained neural networks, useful information when building classifiers or other
relevance aggregation helped identify incorrect rules or predictors” (Bengio, Courville, & Vincent, 2013). From the
machine bias”. The methods of knowledge aggregation outset of the raw data, deep learning methods compute
could be deduced based on metadata, ontology-based, the representation at each layer into a representation at
associated data, social tag and cluster analysis, and a higher and slightly more abstract level (LeCun, Bengio,
knowmetrics. We present meaning and characteristics & Hinton, 2015). Deep learning model Word2Vector and
focusing on each method (Table 1). Wide & Deep learning are used in the empirical study.
Therefore, scholars and research institutions focus Recently, deep learning Word2Vector model has
on the network knowledge aggregation implementation. achieved remarkable success in various text mining
In practice, it has also been promoted, but there are problems such as sentiment analysis, knowledge
still some deficiencies in the current research. From classification, spam filtering, document summarization,
Table 1 in terms of research methods, deep aggregation and web mining. Word2Vector Model is proposed on
of concepts and content at the semantic level of user “Efficient Estimation of Word Representations in Vector
source knowledge cannot be implemented, and the deep Space” (Mikolov, Chen , Corrado, & Dean,2013, p. 4). At the
semantic information cannot be revealed. The sampling same time, Google Inc. provides that Google has developed
group is small and lacks the ability of transforming an efficient tool to implement the algorithm, “an efficient
shallow features of knowledge into deep features. These implementation of the continuous” word vectors, which
deficiencies provide research space for the knowledge is called “Word2Vector” (Introduction, 2013). Many NLP
aggregation in academic virtual community. We use deep applications are promoted and simplified by Word2Vector
learning to overcome the above barriers. model in a critical way (Mikolov et al. ,2013, p. 7). “Many
current NLP systems and techniques treat words as atomic
units - there is no notion of similarity between words,
2.2 Deep Learning as these are represented as indices in a vocabulary”,
and the choices of NLP are simple, robust, and it is
Deep learning is a subset of machine learning in which observed that “simple models trained on huge amounts
multi-layered neural networks—“learn” features from of data outperform complex systems trained on less data”
substantial amounts of data. Within each layer of the neural (Mikolov et al., 2013, p. 1). Semantic relationships are the
network, deep learning algorithms perform calculations, improvement of Word2Vector existing in NPL applications,
make predictions repeatedly, and gradually improve the including machine translation, information retrieval,
accuracy of the outcome over time – all without human and question answering systems. Semantic relationships
intervention (Mueller & Massaron, 2019). Deep learning are beneficial to invent the future of NPL applications
is a nonlinear combination of multi-layer representation (Mikolov et al., 2013, p. 5). Also, the future applications can
An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 375
Figure 1. The CBOW architecture model and the Skip-gram architecture model(Mikolov et al., 2013, p.5).
representations can delete less than the certain frequent 3.1.2 Wide & Deep Learning Model
words and prioritize the evaluation of vocabulary size
to improve the efficiency of the softmax normalization Wide & Deep learning model starts designing to facilitate
(Mikolov et al. ,2013, p. 2–3). with recommend order, but the concept of generalization
We utilize the Skip-gram architecture model in and memorization is preserved to learn. Wide & Deep
the empirical study. In order to improve the training learning consists of two sections. One is wide. The other
efficiency of the model, the strategies of deleting hidden is deep. “Wide learning is a generalized linear model,
layer, negative sampling and hierachy softmax, deleting such as logistic regression or linear regression” (Bastani,
words less than a certain word frequency, and optimizing Asgari, & Namavari, 2019, p. 212). Memorizations are
word weight are adopted in the process of word training. valuable to a wide learning model in the additional
Through the model training, the words are mapped to interaction term. “Memorization is defined as learning the
the abstract high-dimensional vector space. Each word is frequent interactions of features from the historical data”
represented by a dense semantic vector, and the semantic (Bastani et al., 2019, p. 210).
computation can be carried out between the vectors, which Deep learning refers to deep neural networks.
can be used in the fields of similar word computation, Deep neural networks can generalize to unseen feature
word clustering, and semantic computation analysis. In combinations and make the model achieve excellent
this paper, Word2Vec is used to calculate the word vector generalization. Generalization refers to exploiting new
first. After calculating the vectorization representation feature combinations that have never or rarely occurred
of each word, the vector of the word is used to calculate in historical data. Generalization has the adoptability of
corresponding words related to its semantics. We can new samples to add new datasets on the original datasets
capture the semantics of words according to the context by predicting reasonable output in the model training.
of words. This paper innovatively uses the title vector for Generalization has the better adoptability of new samples
semantic analysis, and analyzes the relationship mapping by predicting reasonable output in the model training.
and knowledge association between words and words The model with excellent generalization can learn hidden
at more levels. Word2Vector makes computers “improve rules in data. Hence, for creating a generalized model, it
the existing techniques for estimating the word vectors” uses complex neural network with various hidden layers
(Mikolov et al., 2013, p. 3). to predict the output, including heavy heterogeneity and
mixed consequences.
Wide & Deep learning model can obtain both
generalization and memorization at the same time.
An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 377
Theoretically, Wide & Deep learning model can achieve research items, and so forth. RG has networking skills
accurate classification and wider coverage. with users, such as following users’ accounts, asking
questions, and answering users’ questions. “Statistical
information is available, such as data on readership,
3.2 Building Implementation Process of citations, recommendation counts for research items, and
Knowledge Aggregation System the numbers of questions and answers.” Notable user’s
scientific reputation shows the measurement of an RG
Based on the above theory, we construct the score (Lee, Oh, Dong, Wang, & Burnett, 2019, p. 566–567).
implementation process of knowledge aggregation system As a result, RG resource integration and information
of academic virtual community based on deep learning matching have the characteristics of generalization,
(Figure3). but it is difficult to definitively improve the accuracy of
First, the Word2Vector model is used to train the big information pushes. The purpose of this study is to explore
data that has been cleaned and participled, to calculate the disordered knowledge in academic virtual community
the word vector, to sum or average the vectors of all the for effective semantic disclosure and knowledge link
words in the title, and to obtain the new title vector. and to further realize the deep aggregation of knowledge
Second, cosine distance is used to measure the semantic content. Through big data download and the reveal of
similarity of vectors; it analyzes the semantic mapping semantic vector, RG is selected as an example for the
and knowledge association between words and vectors. implementation of knowledge association and content
Third, K-Means is used to cluster the semantic vectors of depth aggregation. We refine the RG classification
titles to find the knowledge content system after semantic resources for user matching and push information
aggregation. precisely. All work is finished according to Figure 3.
Principal component analysis (PCA) is used to test the
scientific nature and application of aggregation system
by visual display. The operation effect of aggregation 4.1 Data Cleansing and Preparation
system is tested by Wide & Deep learning model, which
can prioritize both generalization and memorization at To verify the scientific and effective organizational
the same time. That is, for any source, new publications structure of knowledge aggregation in network community
according to its title can be automatically carried out based on deep learning, we take the RG (Find and share
clustering navigation. In this way, the whole knowledge ,2008) website as an example; choose a “publications”
aggregation content system can be prioritized by realizing model; search a keyword “Online Communities”; and
user-oriented knowledge accurate push. use Python 3.6 to run 100,000 pieces of data, including
authors, titles, abstract, etc. Then, the 100,000 pieces of
data are cleaned and participled by Python. Similarly,
4 Empirical Study meaningless words and punctuation marks are filtered to
ensure the high purity of the corpus and a better effect of
“RG is the most popular academic social networking the word vector. Therefore, preprocessing could improve
site (ASNS)” in the world and serves global scientific the effect of word vector when using corpus to train word
research users. After users register on RG, a profile vector.
page shows users’ information, such as brief biography,
378 Liangfeng Qian, Shengli Deng
Python Model
100000 Clean Word2Vector Word
ResearchGate
pieces of Model Vector
information Participle Training
Knowledge
Vector Association
Sum
Concept
High Extraction Personalized Precise
Semantic Frequency Aggregation Content
Cluster 内容聚合体系 Recommendation
Keywords System
Association Aggregation
Extraction
Figure 3. The implementation process of the knowledge aggregation system in academic virtual community based on deep learning.
4.2 Data Analysis word vectors. In the meantime, Thomas Mikolov states
that “Word2Vec is a well-known word”. It maps words
Word2Vector is used to reveal semantic vector. Through to a vector representation space and transforms the
the training of Word2Vector, word vectors make full use relationship between words and “surrounding words”
of the contextual information of words and express good into a dense vector. For similar words, the corresponding
semantic features. The information of each dimension in word vectors are similar. Using word vectors can solve
the vector emphasizes the semantic information. We reveal many fields of machine learning, such as computer
the semantic information through the 50 dimensions vision, speech recognition, and NLP (KLUNGPORNKUN &
of words. Semantic similarity is calculated by cosine VATEEKUL ,2018, p. 121).
distance between word vectors. Therefore, the word vector There are two kinds of methods obtaining a set of word
and semantic similarity based on Word2Vector can reveal vectors, which are statistical method and language-based
the semantic vocabulary and the correlation of words model. The more famous method has neural network
from multi-dimensional space. Furthermore, it reveals the language model (NNLM), Skip-gram, CBOW, etc. We use
knowledge association between words and titles. Word2Vector to obtain the characteristics of the words,
which depend on large-scale corpus to obtain accurate
word vectors. In comparison, we collect titles of 100,000
4.2.1 Calculation of the Word Vector Based on retrieval data, clean and participle each title, concatenate
Word2Vector all the titles after processing, and form a word sequence as
the input data of the Word2Vector. The main parameters
“A neural network is constructed, where its input is a are that the word dimension takes 50 dimensions (i.e., 50
sequence of word embedding vectors generated from dimensions per word vector), the context window takes 5
Convolutional Neural Networks (CNN)”. Afterward, words, the words appear at least 10 times, and the number
the researchers mention different training models of of iterative model is 10 times. The result of the calculation
An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 379
Table 2
Examples of Word Vectors
Table 3
The Exhibition of Word Clustering Results
more information and extracts the semantic relationship
between words easily. Therefore, we further construct the
Line number Represent words Norm title vector and use words with conceptual meaning in the
title to analyze the semantic meaning of the title vector
1 Guinea 4.361
(Zhan & Dahal, 2017, p. 4). “Conceptual meaning words”
2 Lanka 4.2377 are selected according to the norm length of the word
3 Sectional 4.2046 vector in each category, and some words with the largest
norm length are selected, namely representative words.
4 Congo 4.1324
So, “conceptual meaning title” is selected according to the
5 Torres 4.1199 norm of the title vector in each category, and some titles
6 Strait 4.1101 with the largest norm are selected.
7 Columbia 4.0069
8 Islander 3.9934
4.2.2 Semantic Analysis of Title Vector
9 Leone 3.8981
Table 4
Title Vector Example
(1) Norm Normalization of Word Vectors; (3) Norm Normalization of Title Vectors;
Assume a word vector is ; the norm of The calculation results of (2) are normalized according
this vector is to the method from (1).
We randomly choose six examples and present
1
m 2 the first five-dimensions of title vector. The remaining
norm = ∑x i2 dimensions are omitted (Table 4):
i=1
The word vector after norm normalization is 4.2.3 Measuring Semantic Similarity
in which .
(2) Sum the Vectors of all the Words in the Title; Similarity at semantic level should consider the relations
Assume there are two-word vectors; the between words. Semantic similarity measurement relates
normalized vectors are and to computing the similarity between terms or short text
Hence, the sum of the word vector expressions that carry the same meaning or related
is and the information but are not lexicographically similar (Martinez-
sum of multiple vectors and so on. Gil & Montes, 2013, p. 399–400). Semantic similarity is
An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 381
( ) ( )
vectors are x11 , x12 , , x1m and m x12 , x 22 , , x 2m Then, 3 Online brand communities 0.85512465
the cosine calculation formula is ∑x1i * x i2 . Because of the 4 Materiality of online brand community 0.85083147
substantial number of words, the similarity calculated
i=1
Table 7
Title Clustering Center Vector
Semantic similarity can be calculated between cluster”. We start choosing k in various ways such that we
title vectors. When the number between title semantic want to find “the number of clusters in the data”. Then,
similarity is higher, the similarity of title is closer. centroids are the centers of those k clusters, “which are
Similarity calculation is the basis of clustering. We initialized in some fashion”. K-Means works best with
cluster similar titles into a class by K-Means++ method to the roughly same sized and shapes clusters in dataset.
implement similarity research topic clusters. “Despite the fact that K-Means is guaranteed to converge,
Based on word vector, we carry on the calculation the final cluster configuration to which it is not coveraged
of title vector for some purposes, such as annotating in general unique and depends on the initial centroid
the semantic relationship between words and words locations” (Visualizing K-means ,2014).
more accurately; realizing the dense vector reflection of K-Means++ is an improvement of the K-Means
words concept, relation, attribute and so on; establishing clustering algorithm in data mining. Based on the K-Means
the vector spatial mapping relation between elements; clustering algorithm, it made some improvements in
and obtaining the new title semantic vector. On this choosing the k initial cluster center. The basic idea is
basis, we use K-Means++, “perhaps an even better that the k initial cluster centers should be as far away as
initialization strategy (though not one implemented in the possible. “The exact algorithm is as follows:”
visualization)” (Visualizing K-means, 2014), to cluster the ① “Choose one center uniformly at random among the
semantic vector of title and realize content aggregation data points”.
based on title vector. ② “For each data point x not chosen yet, compute D(x),
the distance between x and the nearest center that has
already been chosen”.
4.2.4 Knowledge Clustering ③ “Choose one new data point at random as a new center,
using a weighted probability distribution where a
According to the idea of “clustering objects”, objects point x is chosen with probability proportional to
with the same feature are clustered into the same class. D(x)”.
“Clustering is an example of unsupervised learning, in ④ “Repeat Steps 2 and 3 until k centers have been
which we work with completely unlabeled data (or in chosen”.
which our data has labels, but we ignore them)” (Grus, ⑤ “Now that the initial centers have been chosen,
2019). Similarly, clustering refers to enabling an algorithm proceed using standard K-Means clustering”(K-
to recognize these clumps of points without help. The means++, 2020).
K-Means algorithm plays a vital role in clustering analysis.
“The K-Means algorithm captures the insight that each Taking the title vector as the input data, K-Means++
point in a cluster should be near to the center of that algorithm is used to cluster it. The number of clustering
An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 383
Table 8
High-frequency Vocabulary of “Brand Community”
categories is set to 10, and the number of iterations of the and norm refers to the norm of title vector. The larger the
model is 100. The clustering center vectors of 10 categories norm is, the higher the correlation is between the title and
are obtained, as shown in Table 7. the topic. At the same time, it indicates that the title is
The clustering center vector is the vector corresponding more representative of the research content of the topic.
to the center point of each category. The clustering center The categories of topics are ensured according to high-
vector is generally obtained by calculating the mean value frequency keywords and representative titles in the same
of the vectors of all points belonging to the category. way.
If the distance between a point and a clustering center
vector is the shortest, the point belongs to this category.
By clustering the title vector, we can get 10 clustering 4.3 Result
center vectors. Each title only belongs to one of the
categories. That is to say, the distance between the title After data analysis, 10 categories are formed at first. But
vectors corresponding to the clustering center vector is by observing the clustering results, it is found that there is
the shortest. The result of K-Means++ is demonstrated in high similarity between some categories. Some categories
Table 7. According to the title clustering center vector, it cover larger and more complete topics while some
is initially clustered into 10 categories, and then further categories cover more professional and one-sided topics.
optimized the clustering results. So, it is difficult to refine each category of research topics
and difficult to operate.
According to the possible hierarchical clustering
4.2.5 Knowledge Aggregation System relationship between these observations, we try to cluster
five catalogues as the first class (large class) and cluster
The knowledge content is clustered according to five catalogues as second class (small class) under each
title semantic vector. After the clustering calculation first-class catalogue. Thus, the knowledge aggregation
is completed, the concept extraction and semantic system is formed. It is necessary to evaluate and optimize
association are carried out according to the high-frequency the clustering effect. After clustering, the center vector of
keywords in knowledge content. Because of the variety of each category is used as the vector of the category. The
expressions of the same or similar words, it is necessary method in the optimization of clustering system is to
for these to be examined and counted by people. Table compare semantic similarity between each category. If
8 shows the high-frequency keywords in the brand the similarity of two categories is higher than a certain
community. The keyword is combined to synonyms, such threshold value, such as 0.8, it is thought that the two
as communities 830.0, community 727.0 as community categories with high similarity need to be merged.
1557; and behavior 365.0, behaviour 101.0 as behavior 466. While computing the similarity of the first-class
Table 8 clearly shows that under the “online community” category, the similarity between most categories is low.
research category with “brand” as the highest frequency The similarity between the catalogue “online health”
keyword, it can be combined with the research category and the catalogue “community intervention” is 0.805.
and representative title and can be determined as “brand Hence, comparing and analyzing “online health” and
community” for this category of research topics. “community intervention” results in the finding that
After determining the research topic represented by a second-class category highly coincides. So, the two
certain category, the first 50 titles with the largest norm categories are merged.
are taken as the representative titles according to the norm After comparing the similarity of all the second-class
of the unorganized title vector, and the research topics are categories under each first-class category in turn, it is
analyzed again. Table 9 shows the first 10 records here found that the similarity between many categories is high.
384 Liangfeng Qian, Shengli Deng
Table 9
“Brand Community” Representative Title Exhibition
Brand 1 Understanding consumer intention to participate in online travel community and effects on 26.793
community consumer intention to purchase travel online and WOM: An integration of innovation diffusion
theory and TAM with trust
2 An empirical study on the relationship between online shopping mall characteristics and 25.6124
consumers repurchase intention – focused on Chinese consumers involved in overseas direct
purchasing
3 Is online consumers’ impulsive buying beneficial for e-commerce companies? An empirical 23.8119
investigation of online consumers’ past impulsive buying behaviors
4 An empirical study of website personalization effect on user’s intention to revisit e-commerce 20.7455
website through cognitive and hedonic experience: Proceedings of ICDMAI 2018 volume 2
5 Digitalisation luxury fashion and “Chineseness”: The influence of the Chinese context for luxury 20.1041
brands and the online luxury consumers experience
6 The effect of electronic word of mouth on brand image and purchase intention: An empirical 20.0951
study in the automobile industry in Iran
7 A study on the effects of the attractiveness and credibility of online 1 personal media 19.3827
broadcasting B.J. on the viewing engagement perceived on media channel, interactivity,
perceived enjoyment, and the user’s responses.
8 Predicting consumer purchase intention on fashion products in online retailer integration of 18.9955
self-determination theory and theory of planned behavior
9 Extending expectancy violations theory to multiplayer online games: the structure and effects of 18.4863
expectations on attitude toward the advertising attitude toward the brand and purchase intent
10 Negative online reviews of popular products: understanding the effects of review proportion and 17.7826
quality on consumers’ attitude and intention to buy
The reason is that these second-class categories belong to RG, as an example. Downloading big data, revealing
the same first-class category. Therefore, high-frequency semantic vector between titles, clustering by knowledge
keyword is directly adopted in the optimization of second- association, and determining the category theme by high-
class categories. frequency words have ensured the scientific nature of
“Online Community” Knowledge Aggregation the aggregation system theoretically and experimentally,
System based on deep learning (Figure 4) consists of and realized knowledge association and the deep content
five first-class categories, which are semantic analysis, aggregation. The empirical result is “online community”
networking communication, online medical treatment, as a big category resource, which can be refined into
online health, and brand community. Therefore, we can five first-class categories and 25 second-class categories
construct the first-class categories system. After the first- (Figure 4). To test the empirical results, the PCA model
class categories are determined, all samples under each is used to visualize the knowledge aggregation system,
first-class category are clustered by the same reason to and the Wide & Deep learning model is used to test the
get a more detailed category, namely the second-class operation effect of the knowledge system.
categories. Five second-class are formed under the five
first-class categories (Figure 4).
4.4.1 Visual Display of Knowledge Aggregation System
Based on PCA
4.4 Testing Results
We must first standardize, then perform PCA, extract
The “online community” knowledge system based on the desired number of Principal Components (PCs), and
deep learning takes the world’s most popular ASNS, finally use those PCs as input features to neural networks
An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 385
text mining
according to the PC score. Based on the principle of
community question
answering
PCA dimensionality reduction, we use PCA to visualize
semantic analysis
social network analysis
the scientific and reasonable nature of the “online
community” knowledge content aggregation system
based on deep learning. At the same time, this step can
user generated content
visualizing analyzing verify the reliability of the training samples based on Wide
online monitoring & Deep learning classification model.
mobile communication
The effect of classification model depends on the
scientific characteristics and complexity of the model and
network communication
algorithm model
the quality of training data. Traditional clustering model
online communities wireless
telecommunication
requires artificial tagging training data, but artificial
optical communication
tagging is slow and expensive. Therefore, Wide & Deep
model is used for automatic generation of training data
when it chooses training data. The specific methods are
health disaster
customer relationship
management
To verify the scientific of the knowledge system, 15,358
perceived and brand training samples generated by the above methods in five
loyalty
brand community
first-class categories are compressed from 50-dimensional
participation intention
space to 2-dimensional space by PCA model in spark
knowledge sharing ml. After it is compressed to 2-dimensional space, 500
interpersonal trust
training samples are randomly chosen and covering 5
first-class categories. The scatter diagram is visualized by
matplotlib.pyplot, as shown below.
Figure 4. “Online community” knowledge aggregation system.
In Figure 5, stars represent semantic analysis, squares
represent network communication, pentagons represent
(known as neural networks with feature extractions). PCA online medical treatment, “+” represent online health,
is a useful and powerful tool to help us work with high- and triangles represent brand community. Due to the
dimensional problems (many variables) and visualize compression from 50-dimensional space to 2-dimensional
data. Executing PCA is simple with scikit-learn to fit the space, a small spatial error is produced. It is displayed that
model and apply the transformation for data analysis. PCA the sample of five first-class categories is distinguished
is sometimes referred as a feature extraction method based perceptibly.
on many correlated input features to a predictive model Some indicators such as AUC, recall rate, precision,
(Yurko, 2020). “PCA is used widely in dimensionality etc. are applicable in the binary classification model when
reduction” (Dimensionality Reduction, n.d.), especially training with Wide & Deep model in TensorFlow. But in the
the visual exhibition of multi-dimensional data. If multi- multi-classification model, only the accuracy rate as the
dimensional data often encounters dimension greater model calculation results is given. Therefore, this paper
than 3, it cannot be visualized. However, specifying the only lists the calculation results of the model accuracy.
number of components is important to the PCs scores. The calculation results of the model accuracy are
The first two PCs are focused on visualization purposes listed as follows:
386 Liangfeng Qian, Shengli Deng
Table 10
Title First-class Category Automatic Cluster Sample Exhibition
Using stock prices as ground truth in sentiment analysis to generate profitable trading signals 0.907 Semantic analysis
Application of text mining techniques to the analysis of discourse in eWOM communications from a 0.976 Semantic analysis
gender perspective
Creating an Arabic dialect text corpus by exploring Twitter, Facebook, and online newspapers 0.926 Semantic analysis
Table 11
Title Second-class Category Automatic Cluster Sample Exhibition
Utilizing text mining and Kansei engineering to support data-driven design automation at conceptual 0.991 Text mining
design stage
What affects patients’ online decisions: An empirical study of online appointment service based on text 0.991 Text mining
mining: International conference ICSH 2018 Wuhan, China, July 1–3 2018 proceedings
Optimizing cost for geo-distributed storage systems in online social networks 0.902 Social network
analysis
Vulnerability of D2D communications from interconnected social networks 0.933 Social network
analysis
References Liu, J., Zhang, H., & Liu, Z. (2020). Research on online learning
resource recommendation method based on wide & deep and
Elmo Model. Journal of Physics. Conference Series, 1437(1),
Abel, F., Marenzi, I., Nejdl, W., & Zerr, S. (2009). Sharing distri- 012015. Retrieved from doi: 10.1088/1742-6596/1437/1/012015
buted resources in LearnWeb2.0. Lecture Notes in Computer Lu, W., Cai, Y., Che, X., & Lu, Y. (2016). Joint semantic similarity
Science, 5794, 154–159. doi: 10.1007/978-3-642-04636-0_16 assessment with raw corpus and structured ontology for seman-
About. (2008). ResearchGate. Retrieved from https://www.research- tic-oriented service discovery. Personal and Ubiquitous Compu-
gate.net/about ting, 20(3), 311–323. doi: 10.1007/s00779-016-0921-0
Bastani, K., Asgari, E., & Namavari, H. (2019). Wide and deep learning Martinez-Gil, J., & Aldana-Montes, J. (2013). Similarity measurement
for peer-to-peer lending. Expert Systems with Applications, 134, using historical Google search patterns. Information Systems
209–224. doi: 10.1016/j.eswa.2019.05.042 Frontiers, 1 Semantic 5(3), 399–410.
Bengio, Y., Courville, A., & Vincent, P. (2013). Representation lear- Mentzas, G. N., Kafentzis, K., & Georgolios, P. (2007). Knowledge ser-
ning: A review and new perspectives. IEEE Transactions on vices on the semantic web. Communications of the ACM, 50(10),
Pattern Analysis and Machine Intelligence, 35(8), 1798–1828. 53–58. doi: 10.1145/1290958.1290962
doi: 10.1109/TPAMI.2013.50 Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). Efficient esti-
Bi, Q., & Liu, J. (2017). Study on the method of aggregation and mation of word representations in vector space, 1-12. Retrieved
service recommendation of digital resource based on domain from https://arxiv.org/pdf/1301.3781.pdf
ontology. Journal of the China Society for Scientific and Techni- Mu, T., Goulermas, J., Korkontzelos, I., & Ananiadou, S. (2016).
cal Information, 36(5), 452–460. Descriptive document clustering via discriminant learning in
Cheng, H., Koc, L., Harmsen, J., Shaked, T., Chandra, T., Aradhye, a co‐embedded space of multilevel similarities. Journal of the
H., … Shah, H. (2016). Wide & deep learning for recommen- Association for Information Science and Technology. 67, 106-
der systems. DLRS 2016: Proceedings of the 1st Workshop 133.
on Deep Learning for Recommender Systems. 7-10. doi: Mueller, J., & Massaron, L. (2019). Deep Learning for dummies (1st
10.1145/2988450.2988454 ed). For Dummies: Hoboken, NJ, United States.
Dimensionality reduction - rdd-based API. (n.d.). Apache Spark 3.1.1. Tarko, V., & Aligica, P. (2011). From “broad studies” to internet-based
Retrieved from https://spark.apache.org/docs/latest/mllib- “Expert knowledge aggregation.” notes on the methodology
dimensionality-reduction and technology of knowledge integration. Futures, 43(9), 986–
Find and share. (2008). ResearchGate. Retrieved from https://www. 995. doi: 10.1016/j.futures.2011.06.008
researchgate.net/ Visualizing k-means clustering. (2014, January 19). Naftali Harris.
Grisci, B., Krause, M., & Dorn, M. (2021). Relevance aggregation Retrieved from https://www.naftaliharris.com/blog/visualizing-
for neural networks interpretability and knowledge discovery k-means-clustering/
on tabular data. Information Sciences, 559, 111– 129. doi: Yi, M., Mao, J., Deng, W., & Cao, G. (2014). Evolution of know-
10.1016/j.ins.2021.01.052 ledge push network based on social network in social tagging
Grus, J. (2019). Clustering. Data Science from Scratch (2nd ed.). system. Journal of Library Science in China, 40(02), 50–66.
Sebastopol , CA: O’Reilly Media, Inc. Yilmaz, S., & Toklu, S. (2020). A deep learning analysis on question
Introduction. (2013) Word2vector. Google code archive. Retrieved classification task using word2vec representations. Neural
from https://code.google.com/archive/p/word2vec/ Computing & Applications, 32(7), 2909–2928. doi: 10.1007/
Kim, M., Lee, S., & Kim, J. (2020). A wide & deep learning sharing s00521-020-04725-w
input data for regression analysis. IEEE International Confe- Yin, J., & Wang, J. (2014) A dirichlet multinomial mixture model-based
rence on Big Data and Smart Computing (BigComp), 8-12. doi: approach for short text clustering. Association for Computing
10.1109/BigComp48618.2020.0-108 Machinery Digital Library, 233-242.
Klungpornkun, M., & Vateekul, P. (2018). Hierarchical text categori- Yurko, J., Dr (Presenter). (2020, October 9). Principal components
zation using level based neural networks of word embedding analysis (PCA). Lecture presented at University of Pittsburgh,
sequences with sharing layer information. [WJST]. Walailak Pittsburgh, PA, United States.
Journal of Science and Technology, 16(2), 121–131. doi: Zhan, J., & Dahal, B. (2017). Using deep learning for short text under-
10.48048/wjst.2019.4145 standing. Journal of Big Data, 4(1), 1–15. doi: 10.1186/s40537-
K-means++. (November 26, 2020). Wikipedia. Retrieved from http:// 017-0095-2
en.wikipedia.org/wiki/K-means%2B%2B
LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep lear-
ning. Nature, 521(7553), 436–444. doi: 10.1038/nature14539
Lee, J., Oh, S., Dong, H., Wang, F., & Burnett, G. (2019). Motivations for
self‐archiving on an academic social networking site: A study on
Researchgate. Journal of the Association for Information Science
and Technology, 70(6), 563–574. doi: 10.1002/asi.24138
Li, D., Lebo, T., Erickson, J., DiFranzo, D., Williams, G., Li, X., . . .
Hendler, J. (2011). TWC LOGD: A portal for linked open govern-
ment data ecosystems. Journal of Web Semantics, 9(3), 325–
333. doi: 10.1016/j.websem.2011.06.002
Li, Y. (2016). A review on knowledge aggregation. Library and Infor-
mation Service, 60(21), 128–136.