You are on page 1of 17

Data and Information Management, 2021; 5(4): 372–388

Research Article Open Access

Liangfeng Qian, Shengli Deng*

An Empirical Study on Knowledge Aggregation


in Academic Virtual Community Based on Deep
Learning
https://doi.org/10.2478/dim-2021-0010
received April 25, 2021; accepted July 15, 2021.
dimensional knowledge release and knowledge exchange.
As a subdivision product of social networking, academic
Abstract: Academic virtual community provides an social network has become an important platform for
environment for users to exchange knowledge, so it research personnel, professional personnel, and technical
gathers a large amount of knowledge resources and personnel to share knowledge and facilitate academic
presents a trend of rapid and disorderly growth. We learn exchange. ResearchGate (RG) is chosen as an example for
how to organize the scattered and disordered knowledge of empirical study. RG was founded in Germany in 2008, is
network community effectively and provide personalized the professional network for scientists and researchers,
service for users. We focus on analyzing the knowledge collaborates with academic researchers in the world,
association among titles in an all-round way based on deep makes progress in academic virtual community, and has
learning, so as to realize effective knowledge aggregation >20 million members to share their own research work
in academic virtual community. We take ResearchGate (About, 2008). Then, as the representative of RG, user-
(RG) “online community” resources as an example and use generated content (UGC) in academic virtual community
Word2Vec model to realize deep knowledge aggregation. is growing rapidly, and the knowledge accumulation is
Then, principal component analysis (PCA) is used to verify quite considerable. Subject categories and an index of
its scientificity, and Wide & Deep learning model is used mutual citation by authors are used to organize and match
to verify its running effect. The empirical results show that knowledge in RG. Knowledge organization and matching
the knowledge aggregation system of “online community” system are simple. In general, knowledge push and
works well and has scientific rationality. matching correlation are finished according to the interest
and research direction chosen by users. This results in
Keywords: knowledge aggregation, academic virtual a situation wherein the accuracy of push information
community, deep learning, Word2Vec, Wide & Deep is low, and the matching of correlation degree is broad.
learning Therefore, the knowledge integration of academic virtual
community presents the features of less dimension and
shallow level. The lacunae in effective guidance for users’

1 Background knowledge utilization in academic virtual community


restrain knowledge exchange and knowledge innovation.
“Compared with some relative concept,” we think
With the development of Web2.0 technology and computer
“knowledge aggregation is a new direction of knowledge
technology such as big data and cloud computing,
organization, and it contributes to the realization of
academic virtual community has changed the traditional
knowledge service based on user demand.” Knowledge
way of work, life, and education. Users have transitioned
acquisition, knowledge recommendation, and knowledge
from single form of knowledge acquisition to multi-
discovery are the elements of knowledge aggregation in
the big data environment (Li, 2016, p. 128). The concept
of “knowledge aggregation” in the field of knowledge
*Corresponding author: Shengli Deng, Center for Studies of
organization refers to the use of data mining, knowledge
Information Resources, Wuhan University, Wuhan, China, Email:
victorydc@sina.com
mapping, deep learning, artificial intelligence, and
Liangfeng Qian, School of Computing and Information, University of other technologies to describe the external and internal
Pittsburgh, Pennsylvania, USA, Email: Liq12@pitt.edu characteristics of knowledge elements, integrate the

Open Access. © 2021 Liangfeng Qian, Shengli Deng, published by Sciendo. This work is licensed under the Creative Commons
Attribution-NonCommercial-NoDerivatives 3.0 License.
 An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 373

disordered and scattered knowledge elements through 2.1 Traditional Methods


given organization rules, promote the association and
reorganization of multi-source heterogeneous knowledge Scholars start exploring the new direction of knowledge
elements, and devote attention to the deep semantic aggregation from different dimensions to implement
analysis and integration of knowledge content. the ways of knowledge aggregation in academic virtual
The deep and effective knowledge aggregation community, such as ontology-based, metadata, associated
system based on semantic association relies on artificial data, social tag and cluster analysis, and knowmetrics.
intelligence to provide efficient, intelligent, and For example, Mentzas, Kafentzis, and Georgolios (2007)
semantically relevant computer processing technology. mentioned that the concept of Web-based knowledge
Deep learning is an emerging research direction in the AI service can propose an ontology-based framework of
field and realizes highly efficient treatment of input data knowledge aggregation to satisfy the software company
by imitating the human brain. Deep learning is used to by expanding the knowledge management (KM) concept
extract features from input data efficiently, and it can even externally. Abel, Marenzi,, Nejdl, and Zerr. (2009)
extract more abstract features from data, thus enabling reported that the success of recent Web 2.0 platforms
identification and getting more essential features of data. shows that knowledge information aggregation strategy
In this paper, deep learning model Word2Vec is used to is based on tags and metadata for sharing flexible/
annotate the knowledge structure of multi-dimensional relevant learning resources of social networks. Tarko
space (50-dimensional vector representation), which and Aligica (2011) designed “aggregation systems that
can improve the quality of feature extraction and better rely either on meta-experts or on computer algorithms,”
understand the sentences in natural language processing and they explored “the possible of using them for setting
(NLP). In the meantime, Wide & Deep learning is used up virtual thinking tanks for foresight studies”. “By
to construct the classification model of titles with good trans parenting biological evolution theory”, a periodic
memorizations and generalization. The goal of the evolution model is put up “for knowledge push network
paper is to design scientific approach to realize the based on social network under the action of multiple
deep knowledge aggregation in virtual community. The dynamic mechanisms. Then, the whole evolution process
contribution made by this paper is that deep learning of knowledge push based on latent social network is
is used not only to overcome the semantic disclosure of observed and analyzed empirically by using network
knowledge and the lack of grammatical information in structure methods” (Yi., Mao, Deng, & Cao 2014, p. 50).
the network community, but also to express grammatical Li et al. (2011) introduced the TWC(The Tetherless World
information more accurately. Therefore, it integrates Constellation) LOGD(Linked Open Government Data )
the shallow features of knowledge into deep feature Portal to use linked datasets for large-scaling distributed
concepts and knowledge contents, which can express data integration, collaborate data manipulation, and trans
the concepts at the semantic level and reveal the multi- parenting data consumption. Furthermore, Yin & Wang
hidden contents. We emphasize that the semantic (2014) “propose a collapsed Gibbs Sampling algorithm
technology and its related algorithms based on neural for the Dirichlet Multinomial Mixture model for short
network are used to reveal the knowledge association and text clustering (abbr. to GSDMM),” and they “find that
multi-level complex knowledge interaction in academic GSDMM can infer the number of clusters automatically”
virtual community, explore the knowledge disclosure and and “can obtain the representative words of each cluster”.
content aggregation of academic virtual community based Mu, Goulermas, Korkontzelos, and Ananiadou (2016)
on semantic association, and realize the deep aggregation “propose a novel descriptive clustering framework, refer
standardization of knowledge resources and automatic to as CEDL”, and “it discovers an approximate cluster
processing from both theory and practice. structure of documents in the common space” to improve
the effect of knowledge aggregation in academic virtual
community. Bi & Liu (2017) “put forward the approaches
2 Related Works of digital literature resources ontology aggregation and
service recommendation, based on HowNet and literature
This paper studies the means to realize the deep knowledge resources, using a clustering analysis, semantic similarity
aggregation in virtual community. In this section, the computation, collaborative filtering algorithm, and other
existing knowledge aggregation methods and the existing methods”. Based on knowledge aggregation from different
deep learning models that are applied in this paper are scholars, Grisci, Krause, and Dorn. (2021) pointed out,
reviewed. “we present relevance aggregation, an algorithm that
374 Liangfeng Qian, Shengli Deng

Table 1
Comparison of Knowledge Aggregation Methods

Method Meaning Characteristics

Metadata Describe the property of data and realize the unified Simple and easy to use, Strong
integration of heterogeneous knowledge resources standardization, Weak semantics

Ontology-based Formal description of concept system to improve the Strong standardization, Formalized and
machine-readable and understandable data conceptualized, Semantic relevance

Associated data Naming network objects with uniform resource identifier Reveal the semantic meaning and relationship
and data publishing and resource association through HTTP of information to a certain extent
protocol

Social tag and cluster Simple and easy to use, strong freedom, business Poor standardization, Loose structure and
analysis collaboration fuzzy semantics

Knowmetrics Large amount of data processing, multi-dimensional and Weak semantics, Relying on auxiliary tools and
visualization methods

combines the relevance computed from several samples learning methods. Representation learning is “learning
as learned by a neural network and generates scores for representations of the data that make it easier to extract
each input feature. For poorly trained neural networks, useful information when building classifiers or other
relevance aggregation helped identify incorrect rules or predictors” (Bengio, Courville, & Vincent, 2013). From the
machine bias”. The methods of knowledge aggregation outset of the raw data, deep learning methods compute
could be deduced based on metadata, ontology-based, the representation at each layer into a representation at
associated data, social tag and cluster analysis, and a higher and slightly more abstract level (LeCun, Bengio,
knowmetrics. We present meaning and characteristics & Hinton, 2015). Deep learning model Word2Vector and
focusing on each method (Table 1). Wide & Deep learning are used in the empirical study.
Therefore, scholars and research institutions focus Recently, deep learning Word2Vector model has
on the network knowledge aggregation implementation. achieved remarkable success in various text mining
In practice, it has also been promoted, but there are problems such as sentiment analysis, knowledge
still some deficiencies in the current research. From classification, spam filtering, document summarization,
Table 1 in terms of research methods, deep aggregation and web mining. Word2Vector Model is proposed on
of concepts and content at the semantic level of user “Efficient Estimation of Word Representations in Vector
source knowledge cannot be implemented, and the deep Space” (Mikolov, Chen , Corrado, & Dean,2013, p. 4). At the
semantic information cannot be revealed. The sampling same time, Google Inc. provides that Google has developed
group is small and lacks the ability of transforming an efficient tool to implement the algorithm, “an efficient
shallow features of knowledge into deep features. These implementation of the continuous” word vectors, which
deficiencies provide research space for the knowledge is called “Word2Vector” (Introduction, 2013). Many NLP
aggregation in academic virtual community. We use deep applications are promoted and simplified by Word2Vector
learning to overcome the above barriers. model in a critical way (Mikolov et al. ,2013, p. 7). “Many
current NLP systems and techniques treat words as atomic
units - there is no notion of similarity between words,
2.2 Deep Learning as these are represented as indices in a vocabulary”,
and the choices of NLP are simple, robust, and it is
Deep learning is a subset of machine learning in which observed that “simple models trained on huge amounts
multi-layered neural networks—“learn” features from of data outperform complex systems trained on less data”
substantial amounts of data. Within each layer of the neural (Mikolov et al., 2013, p. 1). Semantic relationships are the
network, deep learning algorithms perform calculations, improvement of Word2Vector existing in NPL applications,
make predictions repeatedly, and gradually improve the including machine translation, information retrieval,
accuracy of the outcome over time – all without human and question answering systems. Semantic relationships
intervention (Mueller & Massaron, 2019). Deep learning are beneficial to invent the future of NPL applications
is a nonlinear combination of multi-layer representation (Mikolov et al., 2013, p. 5). Also, the future applications can
 An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 375

develop high-quality Word2Vector significantly (Mikolov 3 A Design Science Approach to


et al., 2013, p. 10). Some deep learning architectures
are utilized in a question classification task in a highly
Building Knowledge Aggregation
inflectional language, namely Turkish, which uses an System Based on Deep Learning
agglutinative-language–based word structure. They have
built word embeddings using the Word2Vec method with The purpose of knowledge aggregation is to build a
a continuous bag-of-words (CBOW) and skip-gram models multi-dimensional, multi-perspective, multi-granularity
with different vector sizes on a large corpus composed of knowledge association system to provide a strong
user questions (Yilmaz & Toklu ,2020, p. 2909). guarantee for user-oriented knowledge services.
Wide & Deep learning is proposed by Google Play, a Therefore, knowledge aggregation focuses on the
commercial mobile app store with a tremendous number knowledge organization based on content association, and
of users and apps, with the published dissertation “Wide excavates the deep learning expression and association of
& Deep learning for Recommender Systems”. Cheng et knowledge resources.
al. mentioned the generalized linear models and deep The problem of knowledge aggregation in academic
neural networks with embeddings and uses jointly virtual community has always been the attention of
trained wide linear model (Cheng et al., 2016). Wide & scholars’ research. We focus on the integration of deep
Deep learning, which combines the wide component learning into the realization of knowledge aggregation,
and deep component, has shown good performance in study how to describe knowledge resources more
recommendation systems. In order to improve the accuracy accurately from multi-dimensional space, and integrate
of online learning platform recommendation for learner massive multi-source heterogeneous data to form an
learning resources and to alleviate the cold start problem, orderly knowledge aggregation system.
an online learning resource recommendation method
based on Wide & Deep and Elmo model are proposed.
Wide & Deep is used to deeply explore the deep features of 3.1 Building Knowledge Aggregation System
learner characteristics and course content features under Based on Deep Learning
the condition of high-dimensional data sparseness. In
addition, for the learner’s text feature, it will use theELMo This paper uses Word2Vec model and Wide & Deep
language model to pre-train the feature vector to improve learning model. In the meantime, the Word2Vec model can
the recommendation accuracy (Liu, Zhang, & Liu., 2020, intelligently interpret the semantic relationship between
p. 1). “The Wide & Deep learning is characterized by the words and sentences. Wide & Deep model can optimize
lack of a research and experimental result on regression generalization ability and memory ability at the same
analysis”. The paper experiments with the application of time. With the utilization of two models, the syntactic
“Wide & Deep learning on regression analysis” and also and semantic features in the text are extracted with high
presents a “new Wide & Deep structure named WDSI”. quality, and the hidden representation of information is
The paper also shows that “the WDSI outperforms a revealed through the contextual semantic relationship as
traditional machine-learning and deep-learning models well as realizing the multi-dimensional aggregation and
in regression analysis” (Kim ,Lee, & Kim., 2020, p. 8). accurate classification of knowledge resources.
Briefly, the shallow features of knowledge are
integrated and abstracted into deep feature concepts
and content, and the semantic-level concept expression 3.1.1 Word2Vector Model
and multi-hidden content disclosure are carried out with
Word2Vector Model to realize the deep aggregation of Word2Vector uses deep learning of recurrent neural
knowledge resources in academic virtual community. Wide network to practice language model. Word2Vector has two
& Deep Learning combining deep neural network with the model architectures for computing vector representations
linear model can be regarded as a content-based method. of words, such as CBOW model and continuous skip-gram
Furthermore, due to the phenomenon that the two parts’ model (Skip-gram). “The CBOW architecture predicts
learning speed is not the same, we need to ensure that the the current word based on” the output of W(t), and “the
two parts are both well-trained. The application field of Skip-gram predicts surrounding words” based on the
the Wide & Deep learning is also expanding. input of W(t) (Mikolov et al., 2013, p. 4–5). In Figure 1, the
word representations are input, projection, hidden and
output layers, and hierarchical SoftMax, and the word
376 Liangfeng Qian, Shengli Deng

Figure 1. The CBOW architecture model and the Skip-gram architecture model(Mikolov et al., 2013, p.5).

representations can delete less than the certain frequent 3.1.2 Wide & Deep Learning Model
words and prioritize the evaluation of vocabulary size
to improve the efficiency of the softmax normalization Wide & Deep learning model starts designing to facilitate
(Mikolov et al. ,2013, p. 2–3). with recommend order, but the concept of generalization
We utilize the Skip-gram architecture model in and memorization is preserved to learn. Wide & Deep
the empirical study. In order to improve the training learning consists of two sections. One is wide. The other
efficiency of the model, the strategies of deleting hidden is deep. “Wide learning is a generalized linear model,
layer, negative sampling and hierachy softmax, deleting such as logistic regression or linear regression” (Bastani,
words less than a certain word frequency, and optimizing Asgari, & Namavari, 2019, p. 212). Memorizations are
word weight are adopted in the process of word training. valuable to a wide learning model in the additional
Through the model training, the words are mapped to interaction term. “Memorization is defined as learning the
the abstract high-dimensional vector space. Each word is frequent interactions of features from the historical data”
represented by a dense semantic vector, and the semantic (Bastani et al., 2019, p. 210).
computation can be carried out between the vectors, which Deep learning refers to deep neural networks.
can be used in the fields of similar word computation, Deep neural networks can generalize to unseen feature
word clustering, and semantic computation analysis. In combinations and make the model achieve excellent
this paper, Word2Vec is used to calculate the word vector generalization. Generalization refers to exploiting new
first. After calculating the vectorization representation feature combinations that have never or rarely occurred
of each word, the vector of the word is used to calculate in historical data. Generalization has the adoptability of
corresponding words related to its semantics. We can new samples to add new datasets on the original datasets
capture the semantics of words according to the context by predicting reasonable output in the model training.
of words. This paper innovatively uses the title vector for Generalization has the better adoptability of new samples
semantic analysis, and analyzes the relationship mapping by predicting reasonable output in the model training.
and knowledge association between words and words The model with excellent generalization can learn hidden
at more levels. Word2Vector makes computers “improve rules in data. Hence, for creating a generalized model, it
the existing techniques for estimating the word vectors” uses complex neural network with various hidden layers
(Mikolov et al., 2013, p. 3). to predict the output, including heavy heterogeneity and
mixed consequences.
Wide & Deep learning model can obtain both
generalization and memorization at the same time.
 An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 377

Figure 2. Wide & Deep learning model(Cheng et al., 2016).

Theoretically, Wide & Deep learning model can achieve research items, and so forth. RG has networking skills
accurate classification and wider coverage. with users, such as following users’ accounts, asking
questions, and answering users’ questions. “Statistical
information is available, such as data on readership,
3.2 Building Implementation Process of citations, recommendation counts for research items, and
Knowledge Aggregation System the numbers of questions and answers.” Notable user’s
scientific reputation shows the measurement of an RG
Based on the above theory, we construct the score (Lee, Oh, Dong, Wang, & Burnett, 2019, p. 566–567).
implementation process of knowledge aggregation system As a result, RG resource integration and information
of academic virtual community based on deep learning matching have the characteristics of generalization,
(Figure3). but it is difficult to definitively improve the accuracy of
First, the Word2Vector model is used to train the big information pushes. The purpose of this study is to explore
data that has been cleaned and participled, to calculate the disordered knowledge in academic virtual community
the word vector, to sum or average the vectors of all the for effective semantic disclosure and knowledge link
words in the title, and to obtain the new title vector. and to further realize the deep aggregation of knowledge
Second, cosine distance is used to measure the semantic content. Through big data download and the reveal of
similarity of vectors; it analyzes the semantic mapping semantic vector, RG is selected as an example for the
and knowledge association between words and vectors. implementation of knowledge association and content
Third, K-Means is used to cluster the semantic vectors of depth aggregation. We refine the RG classification
titles to find the knowledge content system after semantic resources for user matching and push information
aggregation. precisely. All work is finished according to Figure 3.
Principal component analysis (PCA) is used to test the
scientific nature and application of aggregation system
by visual display. The operation effect of aggregation 4.1 Data Cleansing and Preparation
system is tested by Wide & Deep learning model, which
can prioritize both generalization and memorization at To verify the scientific and effective organizational
the same time. That is, for any source, new publications structure of knowledge aggregation in network community
according to its title can be automatically carried out based on deep learning, we take the RG (Find and share
clustering navigation. In this way, the whole knowledge ,2008) website as an example; choose a “publications”
aggregation content system can be prioritized by realizing model; search a keyword “Online Communities”; and
user-oriented knowledge accurate push. use Python 3.6 to run 100,000 pieces of data, including
authors, titles, abstract, etc. Then, the 100,000 pieces of
data are cleaned and participled by Python. Similarly,
4 Empirical Study meaningless words and punctuation marks are filtered to
ensure the high purity of the corpus and a better effect of
“RG is the most popular academic social networking the word vector. Therefore, preprocessing could improve
site (ASNS)” in the world and serves global scientific the effect of word vector when using corpus to train word
research users. After users register on RG, a profile vector.
page shows users’ information, such as brief biography,
378 Liangfeng Qian, Shengli Deng

Python Model
100000 Clean Word2Vector Word
ResearchGate
pieces of Model Vector
information Participle Training
Knowledge
Vector Association
Sum

K-means++ Knowledge cosine


Knowledge Semantic Title
Association Similarity Vector
Mapping
Optimization

Concept
High Extraction Personalized Precise
Semantic Frequency Aggregation Content
Cluster 内容聚合体系 Recommendation
Keywords System
Association Aggregation
Extraction

Operation Effect Scientific


Training Sample
Wide&Deep Test Result
PCA Visual Test
Interactive New Data Learning Model
Test

Figure 3. The implementation process of the knowledge aggregation system in academic virtual community based on deep learning.

4.2 Data Analysis word vectors. In the meantime, Thomas Mikolov states
that “Word2Vec is a well-known word”. It maps words
Word2Vector is used to reveal semantic vector. Through to a vector representation space and transforms the
the training of Word2Vector, word vectors make full use relationship between words and “surrounding words”
of the contextual information of words and express good into a dense vector. For similar words, the corresponding
semantic features. The information of each dimension in word vectors are similar. Using word vectors can solve
the vector emphasizes the semantic information. We reveal many fields of machine learning, such as computer
the semantic information through the 50 dimensions vision, speech recognition, and NLP (KLUNGPORNKUN &
of words. Semantic similarity is calculated by cosine VATEEKUL ,2018, p. 121).
distance between word vectors. Therefore, the word vector There are two kinds of methods obtaining a set of word
and semantic similarity based on Word2Vector can reveal vectors, which are statistical method and language-based
the semantic vocabulary and the correlation of words model. The more famous method has neural network
from multi-dimensional space. Furthermore, it reveals the language model (NNLM), Skip-gram, CBOW, etc. We use
knowledge association between words and titles. Word2Vector to obtain the characteristics of the words,
which depend on large-scale corpus to obtain accurate
word vectors. In comparison, we collect titles of 100,000
4.2.1 Calculation of the Word Vector Based on retrieval data, clean and participle each title, concatenate
Word2Vector all the titles after processing, and form a word sequence as
the input data of the Word2Vector. The main parameters
“A neural network is constructed, where its input is a are that the word dimension takes 50 dimensions (i.e., 50
sequence of word embedding vectors generated from dimensions per word vector), the context window takes 5
Convolutional Neural Networks (CNN)”. Afterward, words, the words appear at least 10 times, and the number
the researchers mention different training models of of iterative model is 10 times. The result of the calculation
 An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 379

Table 2
Examples of Word Vectors

Word Dimension 1 Dimension 2 Dimension 3 Dimension 4 Dimension 5 Dimension


6–50

Online 0.034224428 0.43261373 0.259087235 −0.489322513 −0.007246093 …

Community −0.088755406 −0.020659156 −0.030128725 0.30018574 −0.380684167 …

Communication −0.122577295 0.214154214 0.150527656 0.252177447 −0.154901862 …

Communities −0.097516797 0.141274527 0.143667549 0.444990486 −0.327237487 …

Social −0.210871235 0.029080199 −0.112222128 −0.071832731 −0.195879236 …

Study −0.03604706 −0.101546846 −0.345023751 −0.303048939 −0.532510221 …

Table 3
The Exhibition of Word Clustering Results
more information and extracts the semantic relationship
between words easily. Therefore, we further construct the
Line number Represent words Norm title vector and use words with conceptual meaning in the
title to analyze the semantic meaning of the title vector
1 Guinea 4.361
(Zhan & Dahal, 2017, p. 4). “Conceptual meaning words”
2 Lanka 4.2377 are selected according to the norm length of the word
3 Sectional 4.2046 vector in each category, and some words with the largest
norm length are selected, namely representative words.
4 Congo 4.1324
So, “conceptual meaning title” is selected according to the
5 Torres 4.1199 norm of the title vector in each category, and some titles
6 Strait 4.1101 with the largest norm are selected.
7 Columbia 4.0069

8 Islander 3.9934
4.2.2 Semantic Analysis of Title Vector
9 Leone 3.8981

10 Sierra 3.8625 Word vectors can be calculated by addition and subtraction


to calculate the similarity between sentences. Therefore,
a new semantic vector is obtained by calculating the
is that there are 7988 words with word vector in 100,000 addition or subtraction of the word vector, which can
pieces of data. express the meaning of the combination of words involved
Examples of word vectors are as follows (Table 2), in the calculation. Based on word vector, a new semantic
showing only the first five dimensions, and the remaining vector of title is obtained by calculating the sum and
dimensions are omitted: average of all word vectors in the title.
The purpose of clustering word vectors is to discover In this empirical study, we find that it is the uneven
research topics (Table 3). The norm is used to select the degree of heat and cold due to the inconsistency of the
representative words of each category after clustering. number of words in the training corpus. The norm of the
Table 3 takes some words with the largest norm (from calculated vector is quite different, and the norm of the
big to small) as the representative words, which can words with high frequency is large. In comparison, the
best represent the research content of this category. The norm of the words with low frequency is small. If the
representative words and their norm of the clustering vector is added or averaged directly, the calculated vector
results are shown in the following table: is easily affected by the large norm, and this phenomenon
Because of too fine word granularity, the category of gives rise to what are known as “hot words.” Hot words
research theme is hard to be refined accurately according often appear in subsequent application. Therefore, we
to representative words. The clustering effect is not mention the following calculation methods below to
obvious based on word vector. Due to the composition construct the title vector to eliminate the word hot degree
of words, a title with richer semantic content covers for the effect of constructing title vector.
380 Liangfeng Qian, Shengli Deng

Table 4
Title Vector Example

Title Dimension 1 Dimension 2 Dimension 3 Dimension 4 Dimension 5 Dimension


6–50

A practical guide to the −0.214557097 0.451087397 −0.089244092 −0.105165435 −0.064529825 …


development of an online
course in adapted physical
education

Globalization populism −0.100888215 −0.061636729 −0.163176164 0.05720861 −0.158507724 …


conspiracism

Cohort study −0.094149069 0.058877724 −0.01750199 −0.176174027 −0.150949746 …


evaluating pressure
ulcer management in
clinical practice in the
UK following initial
presentation in the
community costs and
outcomes

A comparative analysis −0.124478352 0.057873331 0.134045295 0.070697202 −0.062715003 …


of metabarcoding and
morphology-based
identification of benthic
communities across
different regional seas

Massive mimo antenna 0.080575986 0.146409381 0.097911457 −0.022636811 −0.045334639 …


array deployment for
airport in air-to-ground
communications

Online brand community −0.031164095 0.178977157 0.376543319 −0.280455168 −0.034272819 …


within the integrated
marketing communication
system when chocolate
becomes seductive like a
person

(1) Norm Normalization of Word Vectors; (3) Norm Normalization of Title Vectors;
Assume a word vector is ; the norm of The calculation results of (2) are normalized according
this vector is to the method from (1).
We randomly choose six examples and present
1
 m 2 the first five-dimensions of title vector. The remaining
norm =  ∑x i2  dimensions are omitted (Table 4):
 i=1 

The word vector after norm normalization is 4.2.3 Measuring Semantic Similarity
in which .
(2) Sum the Vectors of all the Words in the Title; Similarity at semantic level should consider the relations
Assume there are two-word vectors; the between words. Semantic similarity measurement relates
normalized vectors are and to computing the similarity between terms or short text
Hence, the sum of the word vector expressions that carry the same meaning or related
is and the information but are not lexicographically similar (Martinez-
sum of multiple vectors and so on. Gil & Montes, 2013, p. 399–400). Semantic similarity is
 An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 381

one of the key research contents in artificial intelligence Table 5


and NLP. Semantic similarity measurement can increase Word Similarity Example Exhibition
an efficient data matching and supply a precision service.
Rank Similar words of health Similarity
Hence, semantic similarity measurement is widely used in
NLP, information retrieval, and other fields. 1 Healthcare 0.76234574
According to the source of semantic information,
2 Workers 0.71603154
semantic similarity measurement is divided into corpus-
based and knowledge-based. Corpus is an enormous 3 Care 0.69574523

collection of text. Extracted text features in corpus are 4 Interventions 0.6920655


used for constructing feature vectors or statistical models. 5 Mental 0.68639633
“Firstly, the continuous word vectors were trained
6 School-based 0.68160605
from a textual corpus in advance by the neural network
language model in deep learning. Then multiple semantic 7 Preventive 0.68005763
information and relationship information were extracted 8 Visiting 0.65810901
from corpus to augment original vectors and generate
9 Medicare 0.65752329
sense vectors for words. Hence, the semantic similarity
between concepts can be measured by the similarity of 10 Disparities 0.65668325

sense vectors”(Lu, Cai, Che, & Lu, 2016, p. 311–312). We


collect 100,000 pieces of data, construct a huge amount of Table 6
textual corpus, extract text features, and construct feature Title Similarity Example Exhibition
vectors.
Rank Similar title of “online brand community Similarity
The distance between vectors can reflect the similarity
across cultures”
of vectors. The common distance is Euclidean distance,
cosine distance, and so on. 1 Online brand communities loyal to the 0.86801413
community or the brand
We use cosine distance to measure the similarity of
the vectors. Calculation formulas are below: 2 Brand evangelism among online brand 0.86336496
Suppose there are two-word vectors, the normalized community members

( ) ( )
vectors are x11 , x12 , , x1m and m x12 , x 22 , , x 2m Then, 3 Online brand communities 0.85512465
the cosine calculation formula is ∑x1i * x i2 . Because of the 4 Materiality of online brand community 0.85083147
substantial number of words, the similarity calculated
i=1

5 Cultural differences in online community 0.85036181


between each word (title) and all words (title) is too much.
motivations exploring Korean automobile
Based on big data processing software spark, the paper online brand communities(KAOBCs)
calculates the cosine similarity of each word (title) and all and American automobile online brand
words (title) by method of matrix multiplication. Each line communities(AAOBCs)
of the matrix represents a vector of words (titles) when 6 It’s not a shoe it’s a community – 0.84567247
building a matrix. The result of vector multiplication is varumärkesupplevelserpå online brand
cosine similarity after the norm of the vector needs to be communities
normalized. 7 Online brand communities :there is more 0.83576148
We calculate the similarity of each word and all than one way to drive consumers’ online
words first. Then, the similarity is arranged from large to brand identity and interactivity
small, and the first several words are taken as the most 8 Nostalgia in online brand communities 0.83419945
similar words. In the same way, the title similarity can be
9 Online brand community across cultures :a 0.83289393
calculated and analyzed. comparison between the US and Korea
Taking “health” as a similarity example between
10 Online brand community practices and the 0.8318584
words and words, the 10 most relevant words are as
construction of brand legitimacy
follows (Table 5).
Taking “online brand community across cultures” as
a similarity example between titles and titles, the 10 most The range of cosine is (−1, 1). The larger the number is,
relevant titles are shown in the following table (Table 6). the larger the similarity is. In the meantime, 1 represents
completely similarity, but −1 represents no similarity.
382 Liangfeng Qian, Shengli Deng

Table 7
Title Clustering Center Vector

Index Dimension 1 Dimension 2 Dimension 3 Dimension 4 Dimension 5 … Dimension 49 Dimension 50

0 −0.0900 0.2218 0.0692 −0.0649 −0.1430 … −0.0370 0.0318

1 −0.0552 0.2391 0.0992 −0.0776 −0.1130 … −0.1805 0.0725

2 −0.0747 0.1618 0.0813 −0.0415 −0.1946 … −0.0465 −0.0295

3 0.0175 0.2085 0.0767 −0.1440 −0.0342 … −0.0199 −0.0804

4 0.0431 0.1801 0.1236 −0.1602 0.0378 … −0.0791 −0.0500

5 −0.0503 0.1121 0.0518 0.0124 −0.1737 … −0.0669 −0.1022

6 −0.0742 0.1920 0.0417 −0.1041 −0.0618 … −0.1310 −0.2084

7 −0.1194 0.1587 0.1803 −0.2660 −0.1028 … −0.0764 −0.0608

8 −0.0110 0.2546 0.1412 −0.0924 −0.1782 … 0.0270 −0.0244

9 −0.0151 0.1338 0.0822 −0.0194 −0.1065 … 0.0194 0.1752

Semantic similarity can be calculated between cluster”. We start choosing k in various ways such that we
title vectors. When the number between title semantic want to find “the number of clusters in the data”. Then,
similarity is higher, the similarity of title is closer. centroids are the centers of those k clusters, “which are
Similarity calculation is the basis of clustering. We initialized in some fashion”. K-Means works best with
cluster similar titles into a class by K-Means++ method to the roughly same sized and shapes clusters in dataset.
implement similarity research topic clusters. “Despite the fact that K-Means is guaranteed to converge,
Based on word vector, we carry on the calculation the final cluster configuration to which it is not coveraged
of title vector for some purposes, such as annotating in general unique and depends on the initial centroid
the semantic relationship between words and words locations” (Visualizing K-means ,2014).
more accurately; realizing the dense vector reflection of K-Means++ is an improvement of the K-Means
words concept, relation, attribute and so on; establishing clustering algorithm in data mining. Based on the K-Means
the vector spatial mapping relation between elements; clustering algorithm, it made some improvements in
and obtaining the new title semantic vector. On this choosing the k initial cluster center. The basic idea is
basis, we use K-Means++, “perhaps an even better that the k initial cluster centers should be as far away as
initialization strategy (though not one implemented in the possible. “The exact algorithm is as follows:”
visualization)” (Visualizing K-means, 2014), to cluster the ① “Choose one center uniformly at random among the
semantic vector of title and realize content aggregation data points”.
based on title vector. ② “For each data point x not chosen yet, compute D(x),
the distance between x and the nearest center that has
already been chosen”.
4.2.4 Knowledge Clustering ③ “Choose one new data point at random as a new center,
using a weighted probability distribution where a
According to the idea of “clustering objects”, objects point x is chosen with probability proportional to
with the same feature are clustered into the same class. D(x)”.
“Clustering is an example of unsupervised learning, in ④ “Repeat Steps 2 and 3 until k centers have been
which we work with completely unlabeled data (or in chosen”.
which our data has labels, but we ignore them)” (Grus, ⑤ “Now that the initial centers have been chosen,
2019). Similarly, clustering refers to enabling an algorithm proceed using standard K-Means clustering”(K-
to recognize these clumps of points without help. The means++, 2020).
K-Means algorithm plays a vital role in clustering analysis.
“The K-Means algorithm captures the insight that each Taking the title vector as the input data, K-Means++
point in a cluster should be near to the center of that algorithm is used to cluster it. The number of clustering
 An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 383

Table 8
High-frequency Vocabulary of “Brand Community”

Word Online Community Brand Behavior Consumer Loyalty Participation Trust

Freq 4692 1557 895 466 460 459 452 447

Word Intention Knowledge Role Perspective Communication Sharing Interpersonal Study

Freq 423 423 214 417 409 375 366 339

categories is set to 10, and the number of iterations of the and norm refers to the norm of title vector. The larger the
model is 100. The clustering center vectors of 10 categories norm is, the higher the correlation is between the title and
are obtained, as shown in Table 7. the topic. At the same time, it indicates that the title is
The clustering center vector is the vector corresponding more representative of the research content of the topic.
to the center point of each category. The clustering center The categories of topics are ensured according to high-
vector is generally obtained by calculating the mean value frequency keywords and representative titles in the same
of the vectors of all points belonging to the category. way.
If the distance between a point and a clustering center
vector is the shortest, the point belongs to this category.
By clustering the title vector, we can get 10 clustering 4.3 Result
center vectors. Each title only belongs to one of the
categories. That is to say, the distance between the title After data analysis, 10 categories are formed at first. But
vectors corresponding to the clustering center vector is by observing the clustering results, it is found that there is
the shortest. The result of K-Means++ is demonstrated in high similarity between some categories. Some categories
Table 7. According to the title clustering center vector, it cover larger and more complete topics while some
is initially clustered into 10 categories, and then further categories cover more professional and one-sided topics.
optimized the clustering results. So, it is difficult to refine each category of research topics
and difficult to operate.
According to the possible hierarchical clustering
4.2.5 Knowledge Aggregation System relationship between these observations, we try to cluster
five catalogues as the first class (large class) and cluster
The knowledge content is clustered according to five catalogues as second class (small class) under each
title semantic vector. After the clustering calculation first-class catalogue. Thus, the knowledge aggregation
is completed, the concept extraction and semantic system is formed. It is necessary to evaluate and optimize
association are carried out according to the high-frequency the clustering effect. After clustering, the center vector of
keywords in knowledge content. Because of the variety of each category is used as the vector of the category. The
expressions of the same or similar words, it is necessary method in the optimization of clustering system is to
for these to be examined and counted by people. Table compare semantic similarity between each category. If
8 shows the high-frequency keywords in the brand the similarity of two categories is higher than a certain
community. The keyword is combined to synonyms, such threshold value, such as 0.8, it is thought that the two
as communities 830.0, community 727.0 as community categories with high similarity need to be merged.
1557; and behavior 365.0, behaviour 101.0 as behavior 466. While computing the similarity of the first-class
Table 8 clearly shows that under the “online community” category, the similarity between most categories is low.
research category with “brand” as the highest frequency The similarity between the catalogue “online health”
keyword, it can be combined with the research category and the catalogue “community intervention” is 0.805.
and representative title and can be determined as “brand Hence, comparing and analyzing “online health” and
community” for this category of research topics. “community intervention” results in the finding that
After determining the research topic represented by a second-class category highly coincides. So, the two
certain category, the first 50 titles with the largest norm categories are merged.
are taken as the representative titles according to the norm After comparing the similarity of all the second-class
of the unorganized title vector, and the research topics are categories under each first-class category in turn, it is
analyzed again. Table 9 shows the first 10 records here found that the similarity between many categories is high.
384 Liangfeng Qian, Shengli Deng

Table 9
“Brand Community” Representative Title Exhibition

Topic Line number Represent titles Norm

Brand 1 Understanding consumer intention to participate in online travel community and effects on 26.793
community consumer intention to purchase travel online and WOM: An integration of innovation diffusion
theory and TAM with trust

2 An empirical study on the relationship between online shopping mall characteristics and 25.6124
consumers repurchase intention – focused on Chinese consumers involved in overseas direct
purchasing

3 Is online consumers’ impulsive buying beneficial for e-commerce companies? An empirical 23.8119
investigation of online consumers’ past impulsive buying behaviors

4 An empirical study of website personalization effect on user’s intention to revisit e-commerce 20.7455
website through cognitive and hedonic experience: Proceedings of ICDMAI 2018 volume 2

5 Digitalisation luxury fashion and “Chineseness”: The influence of the Chinese context for luxury 20.1041
brands and the online luxury consumers experience

6 The effect of electronic word of mouth on brand image and purchase intention: An empirical 20.0951
study in the automobile industry in Iran

7 A study on the effects of the attractiveness and credibility of online 1 personal media 19.3827
broadcasting B.J. on the viewing engagement perceived on media channel, interactivity,
perceived enjoyment, and the user’s responses.

8 Predicting consumer purchase intention on fashion products in online retailer integration of 18.9955
self-determination theory and theory of planned behavior

9 Extending expectancy violations theory to multiplayer online games: the structure and effects of 18.4863
expectations on attitude toward the advertising attitude toward the brand and purchase intent

10 Negative online reviews of popular products: understanding the effects of review proportion and 17.7826
quality on consumers’ attitude and intention to buy

The reason is that these second-class categories belong to RG, as an example. Downloading big data, revealing
the same first-class category. Therefore, high-frequency semantic vector between titles, clustering by knowledge
keyword is directly adopted in the optimization of second- association, and determining the category theme by high-
class categories. frequency words have ensured the scientific nature of
“Online Community” Knowledge Aggregation the aggregation system theoretically and experimentally,
System based on deep learning (Figure 4) consists of and realized knowledge association and the deep content
five first-class categories, which are semantic analysis, aggregation. The empirical result is “online community”
networking communication, online medical treatment, as a big category resource, which can be refined into
online health, and brand community. Therefore, we can five first-class categories and 25 second-class categories
construct the first-class categories system. After the first- (Figure 4). To test the empirical results, the PCA model
class categories are determined, all samples under each is used to visualize the knowledge aggregation system,
first-class category are clustered by the same reason to and the Wide & Deep learning model is used to test the
get a more detailed category, namely the second-class operation effect of the knowledge system.
categories. Five second-class are formed under the five
first-class categories (Figure 4).
4.4.1 Visual Display of Knowledge Aggregation System
Based on PCA
4.4 Testing Results
We must first standardize, then perform PCA, extract
The “online community” knowledge system based on the desired number of Principal Components (PCs), and
deep learning takes the world’s most popular ASNS, finally use those PCs as input features to neural networks
 An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 385

text mining
according to the PC score. Based on the principle of
community question
answering
PCA dimensionality reduction, we use PCA to visualize
semantic analysis
social network analysis
the scientific and reasonable nature of the “online
community” knowledge content aggregation system
based on deep learning. At the same time, this step can
user generated content

visualizing analyzing verify the reliability of the training samples based on Wide
online monitoring & Deep learning classification model.
mobile communication
The effect of classification model depends on the
scientific characteristics and complexity of the model and
network communication
algorithm model
the quality of training data. Traditional clustering model
online communities wireless
telecommunication
requires artificial tagging training data, but artificial
optical communication
tagging is slow and expensive. Therefore, Wide & Deep
model is used for automatic generation of training data
when it chooses training data. The specific methods are
health disaster

bacterial and microbial as follows:


online medical treatment
hiv/sti
① Construct word vector of all titles (100,000 Pieces of
Data).
② Compute semantic similarity between title vector and
community prevention

online and telehealth cluster center vector (See Section 4.2.3).


③ The most relevant cluster center in ② is chosen,
health education
and the semantic similarity is greater than a certain
mental health
threshold, such as 0.7. The title is used as a reliable
online health
online forum training sample for the cluster center.
④ The training model predicts the title that is not be
community intervention
participated in the training.
healthcare

customer relationship
management
To verify the scientific of the knowledge system, 15,358
perceived and brand training samples generated by the above methods in five
loyalty
brand community
first-class categories are compressed from 50-dimensional
participation intention
space to 2-dimensional space by PCA model in spark
knowledge sharing ml. After it is compressed to 2-dimensional space, 500
interpersonal trust
training samples are randomly chosen and covering 5
first-class categories. The scatter diagram is visualized by
matplotlib.pyplot, as shown below.
Figure 4. “Online community” knowledge aggregation system.
In Figure 5, stars represent semantic analysis, squares
represent network communication, pentagons represent
(known as neural networks with feature extractions). PCA online medical treatment, “+” represent online health,
is a useful and powerful tool to help us work with high- and triangles represent brand community. Due to the
dimensional problems (many variables) and visualize compression from 50-dimensional space to 2-dimensional
data. Executing PCA is simple with scikit-learn to fit the space, a small spatial error is produced. It is displayed that
model and apply the transformation for data analysis. PCA the sample of five first-class categories is distinguished
is sometimes referred as a feature extraction method based perceptibly.
on many correlated input features to a predictive model Some indicators such as AUC, recall rate, precision,
(Yurko, 2020). “PCA is used widely in dimensionality etc. are applicable in the binary classification model when
reduction” (Dimensionality Reduction, n.d.), especially training with Wide & Deep model in TensorFlow. But in the
the visual exhibition of multi-dimensional data. If multi- multi-classification model, only the accuracy rate as the
dimensional data often encounters dimension greater model calculation results is given. Therefore, this paper
than 3, it cannot be visualized. However, specifying the only lists the calculation results of the model accuracy.
number of components is important to the PCs scores. The calculation results of the model accuracy are
The first two PCs are focused on visualization purposes listed as follows:
386 Liangfeng Qian, Shengli Deng

Deep learning model, the word of the title is used as the


feature of the wide part to capture the literal meaning of
the title; the semantic vector of the title is used as the deep
part to capture the semantic meaning of the title.
The model figure is as follows below:
The words in title are represented from word 1 to word
10 in Figure 6. Each title gets at most 10 words. The choice
principle is based on the frequency of each word in the
title from high to low. If >10 words are present in title, the
10 words with the highest frequency are chosen. If <10
words are present in title, the default symbol (usually a
question mark) is used to fill in 10 words. Chosen words
from left to right are used as the input of the wide part of
Figure 5. “Online community” knowledge aggregation system the model.
visualization. The semantic vector of title is represented from dense1
to dense50 in Figure 6, corresponding to the values of
the 1st dimension to the 50th dimension of the semantic
vector of the title in turn.
In Figure 6, multi-class represents predict category.
When the first-class category is clustered, the category
corresponds to five categories in the knowledge
aggregation system. Similarly, when the second-class
category is clustered, it corresponds to the second-class
category under each first-class category – that is, five
5-category models.
Figure 6. Title model based on Wide & Deep learning.
(2) Operation Effect Test of Knowledge Aggregation
System
The accuracy of the “online communities” category As the world’s most popular research social network
is 0.97112024. The accuracy rates of the first class website, 20 million research users share their research
categories are semantic analysis 0.90708065, network results on the platform, and user-generated content is
communication 0.8808989, online medical treatment growing rapidly in RG. Therefore, the efficient knowledge
0.8619744, online health 0.8993711, and brand community aggregation system can quickly cluster multiple elements
0.8224044. Hence, the accuracy of the broad category of latest information from users. We take “semantic
and the first-class category is remarkably high, and analysis” as an example to verify the clustering effect for
the knowledge aggregation system is scientific and any new data(publications). The predict results of several
reasonable. Reliability of training samples of Wide & Deep publications are shown as follows:
learning model is synchronically verified. As a result, the examples in Tables 10 and 11 show
that the fusion of Wide & Deep Learning model and the
“online community” knowledge content aggregation
4.4.2 Operation Effect of Knowledge Aggregation system based on deep learning can cluster publications
System Based on Wide & Deep Learning Model quickly and accurately. In this way, RG online community
publications in the knowledge base can be clustered into
(1) Wide & Deep Learning Model Application five first-class categories and 25 second-class categories.
The application of knowledge aggregation system is to Taking second-class category under “semantic
build intelligent navigation system – that is, for any new analysis” as an example, the prediction results of
date, according to its title, it can automatically cluster the randomly selected publications are as follows:
corresponding research fields. Combined with the idea
of Wide & Deep learning model, we integrate the model
with knowledge aggregation system to construct the
classification model of titles. When building the Wide &
 An Empirical Study on Knowledge Aggregation in Academic Virtual Community Based on Deep Learning 387

Table 10
Title First-class Category Automatic Cluster Sample Exhibition

Title Prob. Category

Using stock prices as ground truth in sentiment analysis to generate profitable trading signals 0.907 Semantic analysis

Application of text mining techniques to the analysis of discourse in eWOM communications from a 0.976 Semantic analysis
gender perspective

Creating an Arabic dialect text corpus by exploring Twitter, Facebook, and online newspapers 0.926 Semantic analysis

N-Gram representations for comment filtering 0.968 Semantic analysis

Table 11
Title Second-class Category Automatic Cluster Sample Exhibition

Title Prob. Category

Utilizing text mining and Kansei engineering to support data-driven design automation at conceptual 0.991 Text mining
design stage

What affects patients’ online decisions: An empirical study of online appointment service based on text 0.991 Text mining
mining: International conference ICSH 2018 Wuhan, China, July 1–3 2018 proceedings

Optimizing cost for geo-distributed storage systems in online social networks 0.902 Social network
analysis

Vulnerability of D2D communications from interconnected social networks 0.933 Social network
analysis

5 Conclusions and Discussions ① The deep learning model Word2Vector carries


deficiencies in its model. The model cannot dispel
The research of knowledge aggregation in academic virtual ambiguity, and each word corresponds to a vector. It is
community based on deep learning takes RG as an example, difficult to learn effective feature vector representation
changes the single method of using Word2Vec to calculate for the features of words with low frequency. The
word vector, converts word vector feature representation Word2Vector model only selects the words in a certain
to higher layer title vector feature representation from window when considering semantic relationship of
raw data, creatively uses this model to calculate more words in the context. The use of context information
semantic title vector, and learns the 50-dimensional vector is limited, and the co-occurrence of word frequency
representation of each title to label the semantic meaning information is not considered from the global
of the title more accurately in order to find the complex perspective.
knowledge association structure in high-dimensional ② Unsupervised word vector training cannot make
data. The knowledge aggregation system is constructed beneficial use of prior information. It is difficult to
empirically by big data. Furthermore, PCA and Wide & generate title vector accurately based on word vector.
Deep learning model are used to verify its scientific and The title vector is obtained by the sum of word vectors.
operation effects. The empirical results show that the When there are too many words in title, the effect of
“online community” knowledge aggregation system is sum may not be good.
scientific and reasonable, and the operation effect is good. ③ The theme of aggregation classification is subjective.
Briefly, the process and method of knowledge aggregation The theme of aggregation classification is clustered
system that is based on deep learning supply new angles according to high-frequency keywords and
and ideas in knowledge aggregation of academic virtual representative title. The statistics of high-frequency
community. keyword synonyms and single/plural word need to be
This research has the deficiencies listed below that finished by people.
require to be addressed in the future research work, so as
to ensure that the line of research elucidated in this paper
is further optimized:
388 Liangfeng Qian, Shengli Deng

References Liu, J., Zhang, H., & Liu, Z. (2020). Research on online learning
resource recommendation method based on wide & deep and
Elmo Model. Journal of Physics. Conference Series, 1437(1),
Abel, F., Marenzi, I., Nejdl, W., & Zerr, S. (2009). Sharing distri- 012015. Retrieved from doi: 10.1088/1742-6596/1437/1/012015
buted resources in LearnWeb2.0. Lecture Notes in Computer Lu, W., Cai, Y., Che, X., & Lu, Y. (2016). Joint semantic similarity
Science, 5794, 154–159. doi: 10.1007/978-3-642-04636-0_16 assessment with raw corpus and structured ontology for seman-
About. (2008). ResearchGate. Retrieved from https://www.research- tic-oriented service discovery. Personal and Ubiquitous Compu-
gate.net/about ting, 20(3), 311–323. doi: 10.1007/s00779-016-0921-0
Bastani, K., Asgari, E., & Namavari, H. (2019). Wide and deep learning Martinez-Gil, J., & Aldana-Montes, J. (2013). Similarity measurement
for peer-to-peer lending. Expert Systems with Applications, 134, using historical Google search patterns. Information Systems
209–224. doi: 10.1016/j.eswa.2019.05.042 Frontiers, 1 Semantic 5(3), 399–410.
Bengio, Y., Courville, A., & Vincent, P. (2013). Representation lear- Mentzas, G. N., Kafentzis, K., & Georgolios, P. (2007). Knowledge ser-
ning: A review and new perspectives. IEEE Transactions on vices on the semantic web. Communications of the ACM, 50(10),
Pattern Analysis and Machine Intelligence, 35(8), 1798–1828. 53–58. doi: 10.1145/1290958.1290962
doi: 10.1109/TPAMI.2013.50 Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). Efficient esti-
Bi, Q., & Liu, J. (2017). Study on the method of aggregation and mation of word representations in vector space, 1-12. Retrieved
service recommendation of digital resource based on domain from https://arxiv.org/pdf/1301.3781.pdf
ontology. Journal of the China Society for Scientific and Techni- Mu, T., Goulermas, J., Korkontzelos, I., & Ananiadou, S. (2016).
cal Information, 36(5), 452–460. Descriptive document clustering via discriminant learning in
Cheng, H., Koc, L., Harmsen, J., Shaked, T., Chandra, T., Aradhye, a co‐embedded space of multilevel similarities. Journal of the
H., … Shah, H. (2016). Wide & deep learning for recommen- Association for Information Science and Technology. 67, 106-
der systems. DLRS 2016: Proceedings of the 1st Workshop 133.
on Deep Learning for Recommender Systems. 7-10. doi: Mueller, J., & Massaron, L. (2019). Deep Learning for dummies (1st
10.1145/2988450.2988454 ed). For Dummies: Hoboken, NJ, United States.
Dimensionality reduction - rdd-based API. (n.d.). Apache Spark 3.1.1. Tarko, V., & Aligica, P. (2011). From “broad studies” to internet-based
Retrieved from https://spark.apache.org/docs/latest/mllib- “Expert knowledge aggregation.” notes on the methodology
dimensionality-reduction and technology of knowledge integration. Futures, 43(9), 986–
Find and share. (2008). ResearchGate. Retrieved from https://www. 995. doi: 10.1016/j.futures.2011.06.008
researchgate.net/ Visualizing k-means clustering. (2014, January 19). Naftali Harris.
Grisci, B., Krause, M., & Dorn, M. (2021). Relevance aggregation Retrieved from https://www.naftaliharris.com/blog/visualizing-
for neural networks interpretability and knowledge discovery k-means-clustering/
on tabular data. Information Sciences, 559, 111– 129. doi: Yi, M., Mao, J., Deng, W., & Cao, G. (2014). Evolution of know-
10.1016/j.ins.2021.01.052 ledge push network based on social network in social tagging
Grus, J. (2019). Clustering. Data Science from Scratch (2nd ed.). system. Journal of Library Science in China, 40(02), 50–66.
Sebastopol ​, CA: O’Reilly Media, Inc. Yilmaz, S., & Toklu, S. (2020). A deep learning analysis on question
Introduction. (2013) Word2vector. Google code archive. Retrieved classification task using word2vec representations. Neural
from https://code.google.com/archive/p/word2vec/ Computing & Applications, 32(7), 2909–2928. doi: 10.1007/
Kim, M., Lee, S., & Kim, J. (2020). A wide & deep learning sharing s00521-020-04725-w
input data for regression analysis. IEEE International Confe- Yin, J., & Wang, J. (2014) A dirichlet multinomial mixture model-based
rence on Big Data and Smart Computing (BigComp), 8-12. doi: approach for short text clustering. Association for Computing
10.1109/BigComp48618.2020.0-108 Machinery Digital Library, 233-242.
Klungpornkun, M., & Vateekul, P. (2018). Hierarchical text categori- Yurko, J., Dr (Presenter). (2020, October 9). Principal components
zation using level based neural networks of word embedding analysis (PCA). Lecture presented at University of Pittsburgh,
sequences with sharing layer information. [WJST]. Walailak Pittsburgh, PA, United States.
Journal of Science and Technology, 16(2), 121–131. doi: Zhan, J., & Dahal, B. (2017). Using deep learning for short text under-
10.48048/wjst.2019.4145 standing. Journal of Big Data, 4(1), 1–15. doi: 10.1186/s40537-
K-means++. (November 26, 2020). Wikipedia. Retrieved from http:// 017-0095-2
en.wikipedia.org/wiki/K-means%2B%2B
LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep lear-
ning. Nature, 521(7553), 436–444. doi: 10.1038/nature14539
Lee, J., Oh, S., Dong, H., Wang, F., & Burnett, G. (2019). Motivations for
self‐archiving on an academic social networking site: A study on
Researchgate. Journal of the Association for Information Science
and Technology, 70(6), 563–574. doi: 10.1002/asi.24138
Li, D., Lebo, T., Erickson, J., DiFranzo, D., Williams, G., Li, X., . . .
Hendler, J. (2011). TWC LOGD: A portal for linked open govern-
ment data ecosystems. Journal of Web Semantics, 9(3), 325–
333. doi: 10.1016/j.websem.2011.06.002
Li, Y. (2016). A review on knowledge aggregation. Library and Infor-
mation Service, 60(21), 128–136.

You might also like