You are on page 1of 7

2016 International Conference on Emerging Technological Trends [ICETT]

Review authenticity verification using supervised


learning and reviewer personality traits
Sonu Liza Christopher Rahulnath H A
Department of Computer Science and Engineering Department of Computer Science and Engineering
TKM College of Engineering TKM College of Engineering
Kollam, India Kollam, India
sonuchristyk90@gmail.com rahulnath@tkmce.ac.in

Abstract— Social media has increasingly promoted users to experience and also guide organizations with marketing and
provide customer feedback on shopping experiences in the form product design strategies.
of online product reviews. Many major ecommerce sites are
collecting revenue from advertisements of products and creating Most review sites are customer centric which make the
better shopping experiences for customers with the help of such presence of fake and unfair competitor placed reviews pose an
reviews. The problem of accurately verifying review authenticity everyday problem. These kinds of untruthful reviews result in
steadily grows and can help in making or breaking the good huge financial gain and grab attention for businesses and
name of a product. Feature engineering performed on the organizations. It has led to the widely known problem of
reviews assist in extracting the useful features that help in opinion spamming seen commonly in the form of fake
identifying actual fake reviews. Supervised machine learning reviews, social network postings such as tweets, and blogs and
algorithms help in classifying reviews into authentic and fake. other forms of deception. These are different forms of illegal
The newly emerging phenomenon of personality prediction has activities which aim at misleading readers by giving
taken hold of social media and this is being employed in the case undeserving positive opinions in order to promote the entities
of reviewer traits which will help identify key personality traits of or provide false negative opinions to some other entities in
fake reviewers. The Big 5 model is used for this which will be order to damage their reputation. The current work deals with
useful in tracking such people through their associated social accurately differentiating fake and authentic product reviews
media accounts.
from online shopping sites predominantly Amazon. Trying to
Keywords—product review, personality trait, social media,
go through such a vast amount of information to understand
opinion mining, feature extraction, review deviation, content the general opinion is impossible for users just by the sheer
similarity, spam detection. volume of data. Every customer is influenced by the recent
market trends and public opinion and hence the need for such
a system. Further, labeling these documents with their
I. INTRODUCTION sentiment would provide a succinct summary to the readers
The onset of social media has brought about massive about the general opinion regarding an entity. Before every
impacts in all spheres of life. From casual comments in social purchase, a customer is involved in analyzing the most recent
networking sites to more serious studies concerning market reviews provided by reliable users of the product and the
trends and analyzing strategies; all which have been possible question rises about which reviews are found to be reliable or
largely due to rising influence of social media data. Data authentic. To determine reliability score for each review is not
mining tasks applied on Big Data has led to knowledge an easy task and these opinions shape the future of the
discovery which has helped analyze market trends and target product. Thus the task of detecting and verifying authentic
customer interests to bring in bigger profits for organizations. reviews or opinions has become critical and significant.
Also people continue to express their interests and opinions
through various social platforms and the collective analyses of II. RELATED WORKS
these texts have also contributed to the growth of Sentiment
analysis. Social media analytics have gained much importance The truthfulness of online product reviews has been widely
and interest among people in present times judging from the discussed in the past. One of the earlier works of Dixit et al.
rapidly growing participation in online social media which has [6] distinguished reviews based on their content as untruthful
risen from 25-43% in many countries. This has motivated reviews, reviews on brands and non reviews. The first
studies which focus on extracting knowledge from opinions category consisting of untruthful reviews is of most concern as
being expressed in social platforms like Twitter, Facebook and they undermine the integrity of the online review system.
also e-commerce sites such as Amazon. One such form of Detection of type 1 review spam is a challenging task as it
social media data are product reviews written by customers attempts to distinguish fake and real reviews by manually
narrating their level of satisfaction with a particular product. reading them.
Organizations such as Amazon are involved in hundreds and Data mining and machine learning techniques together
thousands of electronic transactions every day where the have been effective in detecting fraudulent reviews. A
reviews of customers help new buyers make a better shopping straightforward example of web content mining is opinion

978-1-5090-3751-3/16/$31.00 ©2016 IEEE


2016 International Conference on Emerging Technological Trends [ICETT]

mining which consists of attempting to ascertain the sentiment


(positive or negative) of a text by analyzing its features.
Review authenticity verification also lies in the category of
content mining and in addition utilizes features not directly
linked to the content. These features are constructed from the
online reviews using various different techniques in text
mining and Natural Language Processing (NLP). There are
also features linked with the review author for instance their
frequency pattern and how much a review deviates from
reviews for the same product or service which allows
distinguishing between authentic and fake reviews.
Existing machine learning techniques were not sufficiently
accurate in review authenticity verification, but it was found
these methods are more reliable than manual detection. The
primary issue, which was identified by Abbasi et al. [1], is the
lack of any distinguishing features which can be based on to
act as a definitive guide in the classification process of Fig. 1. Proposed System Architecture
reviews as real or fake. A common method is using the bag of
words approach in which the presence of individual or small
III. PROPOSED METHOD
groups of words is used as the feature set. But studies have
shown that this approach is not sufficient to train a classifier The entire model of review authenticity verification system
with adequate performance. This led to modified methods of is divided into two modules. The first module comprises of the
feature engineering in an effort to extract a more informative Review Authenticity Verification Module which performs
feature set that will improve review authenticity verification. tasks such as pre-processing, feature extraction and machine
There are many previous works that take into account different based classification using SVM. The dataset comprises of
features sets for the current context of review authenticity, reviews extracted from Amazon website of Musical
each utilizing a different variety of machine learning instruments [35]. The customer reviews relate experiences that
techniques. Jindal et al. [10], Li et al. [19] and Mukherjee et each person has endured during the course of post-purchase.
al. [25], used individual words from the review text as the Some reviews are lengthy and dwell into the musical history
features, while Shojaee et al. [31] used syntactic and lexical and describe passion towards music. To reduce the size of data
features. An additional study by Ott et al. [29] used review to be processed, two tasks are done. First stop word removal
characteristic features in addition to unigram and bigram term- removes words which do not convey a sentiment or emotion
frequencies. Features associated with the behavior of the and can be avoided. This help is considerably reducing the
reviewer also directed further investigation as it helps in data size. The next part of the preprocessing process deals
extracting information from multiple texts of the same with POS tagging. Every word in a review is tagged with its
reviewer. Examples of studying spammer behavior include part of speech .The most common parts of speech are nouns,
spotting multiple User IDs for the same author [30] and adjectives, verbs, adverbs etc. After pre-processing the process
identifying groups of spammers by studying their behavioral of feature selection or extraction is carried out. This deals with
footprints [23–25]. Combining review spam detection through extracting essential and crucial features from each review
a review’s features, and spammer detection through analysis which best aid in verifying review authenticity. Previous
of their behavior was found to be a more effective approach works have made use of different features such as activity
for detecting review authenticity than either approach alone. window, review count maximum number of reviews, number
of positive reviews, review deviation, content similarity. From
Before addressing the challenges associated with the already mentioned factors a subset of reliable features is
improving review authenticity verification, we must first selected after study and these observed values are passed to
address collection of data. Data is a major part of any machine SVM for machine based classification. Fig 1 depicts the
learning based model, and while a massive volume of reviews proposed system architecture.
are available on the Internet, collecting and labeling a
sufficient number of them to train a review spam classifier is a The second module deals with personality prediction using
difficult task. An alternative to collecting and labeling data is Big 5 personality model. This module takes as input only
to artificially create review spam datasets by the method of those reviews identified by the output of the classification task
synthetic review spamming. This takes existing truthful as fake reviews. This is done to uncover core personality traits
reviews and builds fake reviews from them. Also review of fake reviewers based on the reviews written by them after
burstiness was studied in [8] which have also been the initial identification of such reviewers. These traits can be
implemented in the proposed system with modifications in utilized to find out the presence of fake reviewers from social
detecting burst periods. media like Facebook and Twitter where personality prediction
is currently being implemented. From each identified fake
review, features are extracted which include average word
count, average length of a word, number of positive and
negative words, number of punctuations, exclamations,
2016 International Conference on Emerging Technological Trends [ICETT]

question mark etc which help in building a grammatical vector • Review deviation
representing each reviewer. The Big Five describes a
personality structure divided into five basic elements known Every review is marked by a reviewer with a review
as OCEAN: Openness; Conscientiousness; Extroversion; rating that denotes the level of satisfaction achieved from
Agreeableness; and Neuroticism. These are the five core using the product. For a product p, the rating is marked
personality traits and the presence of each trait is measured in using a five point scale ranging from a very bad rating of
each reviewer and the most common traits identified. Since a 1 to a very high rating of 5. The maximum deviation
reviewer cannot be analyzed in terms of all five traits in the possible for any product is 4. To calculate the review
model, the presence of only traits like extroversion, deviation for a product by a given reviewer, the average
neuroticism and openness will be taken in this work. rating of all customers except current reviewer is taken
and review deviation is calculated. To normalize the score
this value is divided by a factor of 4.This can be
A. Preprocessing
expressed by the formula defined below. If the average
Preprocessing is the first step in any text mining process deviation is greater than a selected threshold value, then
and plays a very important role. The most common words in the reviewer is marked as likely fake.
text documents are articles, prepositions, and pro-nouns which
do not add any specific meaning to the text. These words are RD(p) = |R(p) - R(avg,
4
p)|
(1)
treated as stop words. The reason for removing stop-words
from a text is because they take up additional processing and • Burst Detection
this can be avoided by the dimensionality space of the review
text being analyzed. Many types of stop word removal There are durations within a product lifetime when
methods are followed which are used to remove stop words there are large number of reviews flooding in. These may
from the data which in this case are the product reviews. The be because of sudden popularity of the product which may
classic method is based on removing stop words obtained from be attributed to advertisements and other seasonal
an already pre-compiled list. Another approach includes promotions or else due to fake reviewer groups who write
removing words with low inverse document frequency (IDF) many reviews in a day on various products. These are
and mutual information. The lower value of mutual known as burst periods and detection of such burst periods
information suggests that the sop-word relates to a low are an indication of likely fake reviews. Given a product
discrimination power hence can be removed from the review which has a set of m reviews, the first step is to divide the
text. The classical method has been adopted in the proposed life span of the product into small sub-intervals or bins.
system which uses a file of stop words and removes every This is done by choosing a proper bin size (e.g. two
occurrence of the word if present in the review text. weeks). Next step is finding the average number of
reviews in each bin. For all those months for which the
The second step of pre-processing uses POS tagging. review count is greater than average, mark flag 1 to denote
Traditional parts of speech are nouns, verbs, adverbs, exceptionally high review count for specific duration and
conjunctions, etc. Part-of-speech taggers typically take a mark others 0. This is done for each year in the lifespan of
sequence of words (i.e. a review) as input, and provide a list of the product. After each month of every year is labeled
tuples as output, where each word is associated with the either 1 or 0, next we select every 2 years and check to see
related tag based on both its definition and its context. The recurring patterns occur like seasonal offers or promotions
tagger used in this work is the Stanford POS tagger which which would explain the higher review count for that
provides an accuracy of 96.7%. Taggers are available in month. If such a repeating pattern is observed within a
various different languages. The English taggers use the Penn span of 2 years, then the burst flag is labeled 0 and if not
Treebank tag set. In the proposed system the tagger takes as 1.This means that every month which does not repeat and
input a file containing review text and annotates each word has a higher review count in a particular year is prone to be
with its corresponding tag. This can be used to find the count a fake review. All reviews for the product during the
of each part of speech in a review. This will later be used in particular month are marked 1 which denotes likely fake.
the personality prediction module. This factor is given lesser weight than other features
because burst flag does not prove all reviews during that
B. Feature Extraction period to be fake and only a subset of such reviews may be
The features are divided into two categories of review and actually fake.
reviewer centric features. Review centric features are those • Review Content Similarity
constructed using information available from a single review
and reviewer centric features adapt a holistic approach taking This feature has been used previously and has proved
into account all of the reviews written by any particular to be useful in identifying fake reviews. Clearly this is a
reviewer along with any additional information regarding the reviewer centric feature as it takes in account all the
reviewer himself. In this case, profile information relating to a reviews written by a reviewer and measures the content
reviewer is not available so we rely only on the information similarity value between each review. Content similarity is
collected from past reviews of a reviewer to build the reviewer a measure which can be found using various methods such
centric features. The 4 basic features used in the proposed as cosine similarity, Jaccard similarity and various other
system are defined as follows. similarity measures. Cosine similarity is selected because
it gives a fast and accurate value for similarity between
2016 International Conference on Emerging Technological Trends [ICETT]

texts. This is done using the term frequency and inverse


document frequency of each term in the review. The
average is calculated across all reviews and cosine
similarity values of the reviewer. For a reviewer having
high content similarity of value 0.5 or higher, the reviews
are marked as likely spam. For a reviewer who has no
more than a single review with a unique reviewer id,
content similarity value is not applicable in such scenario.
• Sentiment Score
Finding the sentiment of a reviewer can help assign a
small score to each review. A person who only writes
very good or very bad reviews can be treated as a fake
reviewer. To find this we use Stanford CoreNLP package
in which each sentence is assigned a sentiment score
ranging from 0 to 5. This feature is also reviewer centric Fig. 2. Big 5 Model
and helps analyze the sentiment of reviews written by a
particular reviewer. Calculating the average score of all
sentences in a review each review is marked with a score
and normalized. A sentiment score above a 0.6 denotes a multi-label problem into a binary problem, while the
very positive review and below 0.3 denotes a very bad classification module performs the actual task of personality
one. Next we proceed by calculating the average classification itself. These modules are detailed in the
sentiment scores of all reviews of a particular reviewer. A following sections.
one-time reviewer cannot be judged with this factor and
A. Pre-processing module
so this lowers its priority because of the limit in universal
applicability to identify fake reviews written by an This module does not work directly with the reviews but
individual who hasn’t written any other reviews before. A instead uses information extracted from the review texts. In its
lower weight is thus assigned to this factor. first module the initial text corpus and user information are
transformed into a set containing some grammatical features
C. Classification using supervised learning highlighted below. User information is not available in the
Machine learning comes in various forms of supervised, current context but can be taken up as future work which can
semi-supervised and unsupervised methods. Earlier works help analysis when a reviewer can be tracked to his social
have been done creating a gold standard dataset in which media accounts in Facebook or Twitter. The information
accuracy and performance of classifiers like Support Vector extracted from the reviews is used to create a set of
Machine was found to be very high. Keeping in mind the lack grammatical category of features. The grammatical category
of authenticity in creating a dataset, the currently in use takes into account information like the number of words used,
dataset of Amazon based product reviews have a predefined number of punctuation marks etc utilized in the text. The
structure. Each review is described using a set of features like meta-attributes extracted are as follows
reviewer name, reviewer ID, product code, review rating,
review text and review date. The data has been modified to a • Average length of text
small extent to facilitate false testing and labels are manually • Average number of words.
assigned to a small subset of reviews. This is done with the • Average number of words longer than six letters.
help of a small training data set and the SVM is trained to
• Average number of unique words.
predict the labels for each review. Supervised learning is
performed using classification by SVM. Other classifiers such • Average number of question marks.
as Bayesian classifiers were earlier used but accuracy and • Average number of exclamation marks.
performance measured were found to be less. The final label • Average number of positive words.
assigned to each review indicates 0 in the case of authentic • Average number of negative words.
review and 1 for every fake review.
Every review is represented as a vector of the above found
IV. PERSONALITY PREDICTION SUBSYSTEM features and is next passed to the transformation module
which deals with individual binary problems; that is, for each
The three basic steps for personality detection comprises of the Big Five personality traits a binary classifier is created
of pre-processing, transformation and classification. The five with two labels which indicate the presence or absence of the
traits of the OCEAN model are highlighted in Fig 2. The pre- trait.
processing module is responsible for extracting a set of
features from the previously identified fake reviews. The B. Transformation module
transformation module handles the task of converting each
In this module, the primary task deals with transformation
of multi label problems into linear single label problems where
2016 International Conference on Emerging Technological Trends [ICETT]

each label will correspond to a personality trait in the Big Five used product reviews from Amazon belonging to the musical
model of a binary problem; that is, a problem in which each instrument category and the features extracted are subjective
review text may or may not contain a personality trait. Let R to this category of products. Each review is represented by a
be a reviewer characterized by the traits extroversion and set of features including reviewer name and ID, product code,
openness. After the transformation, this reviewer now has all user review rating, review date and review text.
labels with extroversion and openness marked with ‘1’ and the All reviews are initially loaded from a file into table and
other three personality traits, agreeableness, conscientiousness each reviewer details are inserted into a separate Reviewer
and neuroticism, marked with ‘0’. In the upcoming stage of Table. Content similarity values between every two reviews of
classification module, each classifier is assigned the role to a particular reviewer are calculated and saved linked to each
determine whether a reviewer has the personality trait. The distinct reviewer. Also against each reviewer the number of
presence of a trait is marked by 1 and absence by 0. Only three reviews written during bursts periods is also noted which
classifiers are used in the present work because as previously comprised the burst ratio feature. Sentiment scores are
explained review texts would not be useful in finding out traits calculated using deep learning approaches. Each sentence of a
like conscientiousness and agreeableness. review is annotated and sentiment calculated and average
score for a review is calculated. This helps calculate the
C. Classification module overall sentiment score of each reviewer. The burst detection
One of the most commonly used methods of is carried out on separate product ids and for each year that the
classification in machine learning is the use of semi- product was written about. Every month of the year is marked
supervised techniques which uses a small number of labeled by 1 or 0 indicating the presence of burst or not. If a burst
data and a large number of unlabelled data. This is ideal in the month reappears every time in two consecutive years, such a
present case as the product reviews do not have a previously scenario is considered to be due to natural reasons like
annotated text labeled with personality traits. Personality seasonal promotions which occur during the same months
prediction has been previously carried out in many social every year. In this case, a small weight is added to this feature
networking sites but as to the best of our knowledge it hasn’t because this does not have to be the case always as fake
been applied on product reviews. As a labeled dataset is not reviews are also present during seasonal times. Every other
available in the present scenario, we are taking up an approach month for the year under observation is marked likely burst
from a previous work in which prediction was performed on a with a flag 1.
set of Tweets. The data set taken for this purpose is
synthetically created by analyzing a similar data set of Tweets Therefore we can treat the spam reviews as
having an approximate length corresponding to the review belonging to the positive class and non-spam reviews as
lengths of the fake reviewers. The data matching the overall belonging to the negative class. A classifier can then be built
length of the review texts was analyzed and a dataset was to separate the two classes of reviews. The values for
created with labeled traits of extroversion, neuroticism and accuracy, precision, recall and f-measure are described in
openness. The grammatical features are studied and a
comparative study is done to create a small labeled set of traits
for each review. This method was adopted because of the lack
of presence of a labeled set for product reviews.

TABLE 1. EXPERIMENTAL VALUES FOR PRECISION, ACCURACY, RECALL AND F-


MEASURE

Dataset Used: Amazon Reviews (Musical Instruments)


Method Used Precision Recall Accuracy F-measure
SVM with Proposed 0.9901 0.7368
0.7 0.778
set of Review Features
SVM with Basic set of 0.9839 0.4285
0.3 0.75
Review Features

Fig. 3. Comparison of accuracy, precision, recall and f-measure


V. EXPERIMENTAL EVALUATION AND RESULTS for proposed and baseline methods
One of the major obstacles towards review authenticity
detection was the evaluation because there is no ground truth Table 1. The labels predicted by the classifiers could be
data of review authenticity that can be used in model building categorized as True Positive (TP), True Negative (TN), False
and model testing. Previously human evaluation has been used Negative (FN) or False Positive (FP). A True Positive
as a means to test the authenticity of reviews but this would
(likewise Negative) refers to an object whose actual class is
fail due to the subjective nature of each evaluator who would
positive (negative) and the classifier predicted it to be positive
possess different tolerance levels even if they are given the
same set of behavior indicators and reviews. We now evaluate (negative). False Negative classes correspond to those objects
the proposed system in Java on an Intel® Core i3 processor whose actual is positive and the classifier predicted them to be
running Windows 8.1 with 4 GB of main memory. We have negative. Similarly, the true label of False Positive class is
negative and the classifier predicted them to be positive.
2016 International Conference on Emerging Technological Trends [ICETT]

highlighted the predominant presence of extroversion in most


fake reviewers followed by openness and still lower levels of
neuroticism (Fig 4). Since the data taken for evaluation was
synthetically created, the personality traits found by this
method can be used as a baseline for further studies in this
area. Personality could be still strengthened by analyzing the
social behavior of reviewers which has not been taken up in
this study. The total time taken to extract each feature and
overall memory consumption for entire system is less.
VI. CONCLUSION
The proposed method for review authenticity verification
shows a significant increase in accuracy and precision from
the baseline method. This is due to the selection of features
Fig. 4. Extent of presence of Extroversion, Openness and Neuroticism in such as burst review detection, sentiment scores, content
identified fake reviewers similarity, review deviation and certain modifications and
weights assigned to each factor which have lead to this
• Precision is defined as the number of true positives improvement. Furthermore personality detection module has
(TP) over the number of true positives plus the been applied for the first time in the case of product reviews
number of false positives (FP): which assess grammatical features of all reviews of a
Precision = TPTP (2) particular reviewer and predict the most common feature of a
+ FP
fake reviewer. The results imply that the most common
• Recall is defined as the number of true positives (TP) personality traits that a fake reviewer is likely to have is
over the number of true positives plus the number of extroversion. This trait makes encourages people to be more
false negatives (FN): active on social media and in sharing their opinions in a very
strong and loud manner. Any fake reviewer would deviate
Recall = TPTP
+ FN
(3) significantly from general public opinion and influence buyers
to make wrong choices in a very convincing manner. All this
• Accuracy is defined as the overall success of the could be accounted to an extrovert who is active, enthusiastic
classifier, i.e. is the average number of correct and fully focused on spreading the wrong word. Future works
predictions divided by the number of labels: can be done focused on linking the various reviewer ids in
product reviews to the social media accounts of these
Accuracy = TP +TN
TP + TN
+ FP + FN
(4) reviewers. As the participation in social media is increasing
every day, performing analysis on social media data would
reveal more behavioral traits of a person. Facebook and
• F-measure is calculated by, Twitter are two platforms where the social behavior of a
person can be studied in depth and sentiment analysis on posts
precision∗ recall
F = 2∗ precision + recall
(5) and comments would shed light in different dimensions.
The supervised learning is performed using SVM based
classifier. The values for accuracy, precision, recall and f-
measure are described in Table 1. A comparison of the values References
obtained using the proposed system against the baseline
method is illustrated in Figure 3. The reviews identified as [1] Abbasi, Ahmed, et al. "Detecting fake websites: the contribution of
those belonging to fake reviewers were passed to the statistical learning theory." Mis Quarterly (2010).
personality prediction module which performed multi [2] Aaron N. Richter , Hamzah Al Najada , Joseph D. Prusa ,Michael
classification to train the classifier and find out the percentage Crawford , “Survey of review spam detection using machine learning
of traits identified from such reviewers. The data set for semi- techniques” , Crawford et al. Journal of Big Data (2015).
supervised classification of personality traits was manually [3] Bandakkanavar, Rekha V., M. Ramesh, and H. Geeta. "A survey on
created using a similar set of Tweets used in one of the prior detection of reviews using sentiment classification of methods."
IJRITCC 2.2 (2014).
works for personality prediction [5]. The data set created
[4] Banerjee, Snehasish, Alton YK Chua, and Jung-Jae Kim. "Using
contained reviews with similar review length as the tweets. supervised learning to classify authentic and fake online reviews."
The corresponding grammatical features were modified to an Proceedings of the 9th International Conference on Ubiquitous
extent as they would vary within the different contexts. Information Management and Communication. ACM, 2015.
[5] Lima, A. C. E., \& De Castro, L. N. (2014). A multi-label, semi-
The created set contains reviews having presence of all supervised classification approach applied to personality prediction in
three traits in varying patterns. The semi-supervised social media. Neural Networks.
classification was then conducted on the fake reviews [6] Dixit, Snehal, and A. J. Agrawal. "Survey on review spam detection."
identified from previous module. Although the number of fake Int J Comput Commun Technol ISSN (PRINT) 4 (2013).
reviews identified was less, the proposed prediction module
2016 International Conference on Emerging Technological Trends [ICETT]

[7] Dave, Kushal, Steve Lawrence, and David M. Pennock. "Mining the [22] Mukherjee, Arjun, et al. Fake review detection: Classification and
peanut gallery: Opinion extraction and semantic classification of product analysis of real and pseudo reviews. UIC-CS-03-2013. Technical
reviews." Proceedings of the 12th international conference on World Report, 2013.
Wide Web. ACM, 2003. [23] Mukherjee, Arjun, et al. "Spotting opinion spammers using behavioral
[8] Fei, Geli, et al. "Exploiting Burstiness in Reviews for Review Spammer footprints." Proceedings of the 19th ACM SIGKDD international
Detection." ICWSM 13 (2013). conference on Knowledge discovery and data mining. ACM, 2013.
[9] Feng, Song, et al. "Distributional Footprints of Deceptive Product [24] Mukherjee, Arjun, et al. "What yelp fake review filter might be doing?."
Reviews." ICWSM 12 (2012). ICWSM. 2013.
[10] Jindal, Nitin, Bing Liu, and Ee-Peng Lim. "Finding unusual review [25] Mukherjee, Arjun, Bing Liu, and Natalie Glance. "Spotting fake
patterns using unexpected rules." Proceedings of the 19th ACM reviewer groups in consumer reviews." Proceedings of the 21st
international conference on Information and knowledge management. international conference on World Wide Web. ACM, 2012.
ACM, 2010. [26] Narayanan, Vivek, Ishan Arora, and Arjun Bhatia. "Fast and accurate
[11] Jindal, Nitin, and Bing Liu. "Opinion spam and analysis." Proceedings sentiment classification using an enhanced Naive Bayes model."
of the 2008 International Conference on Web Search and Data Mining. International Conference on Intelligent Data Engineering and Automated
ACM, 2008. Learning. Springer Berlin Heidelberg, 2013.
[12] Jindal, Nitin, and Bing Liu. "Review spam detection." Proceedings of [27] Pang, Bo, and Lillian Lee. "Opinion mining and sentiment analysis."
the 16th international conference on World Wide Web. ACM, 2007. Foundations and trends in information retrieval (2008).
[13] Judge, Timothy A., et al. "The big five personality traits, general mental [28] Ott, Myle, et al. "Finding deceptive opinion spam by any stretch of the
ability, and career success across the life span." Personnel psychology imagination." Proceedings of the 49th Annual Meeting of the
52.3 (1999). Association for Computational Linguistics: Human Language
[14] Khan, Khairullah, et al. "Mining opinion components from unstructured Technologies-Volume 1. Association for Computational Linguistics,
reviews: A review." Journal of King Saud University-Computer and 2011.
Information Sciences(2014). [29] Ott, Myle, Claire Cardie, and Jeffrey T. Hancock. "Negative Deceptive
[15] Khan, Khairullah, Baharum B. Baharudin, and Aurangzeb Khan. Opinion Spam." HLT-NAACL. 2013.
"Mining opinion from text documents: A survey." 2009 3rd IEEE [30] Qian, Tieyun, and Bing Liu. "Identifying Multiple Userids of the Same
International Conference on Digital Ecosystems and Technologies. Author." EMNLP. 2013.
IEEE, 2009. [31] Shojaee, Somayeh, et al. "Detecting deceptive reviews using lexical and
[16] Laryea, Bryan Nii Lartey, et al. "Web Application for Sentiment syntactic features." 2013 13th International Conference on Intellient
Analysis Using Supervised Machine Learning." International Journal of Systems Design and Applications. IEEE, 2013.
Software Engineering and Its Applications 9.1 (2015) [32] Saulsman, Lisa M., and Andrew C. Page. "The five-factor model and
[17] Lau, Raymond YK, et al. "Text mining and probabilistic language personality disorder empirical literature: A meta-analytic review."
modeling for online review spam detecting." ACM Transactions on Clinical psychology review(2004).
Management Information Systems 2.4 (2011). [33] Virmani, Deepali, Vikrant Malhotra, and Ridhi Tyagi. "Sentiment
[18] Li, Jiwei, et al. "Towards a General Rule for Identifying Deceptive Analysis Using Collaborated Opinion Mining." arXiv preprint arXiv
Opinion Spam." ACL (1). 2014. (2014).
[19] Li, Fangtao, et al. "Learning to identify review spam." IJCAI [34] Xie, Sihong, et al. "Review spam detection via temporal pattern
Proceedings-International Joint Conference on Artificial Intelligence. discovery." Proceedings of the 18th ACM SIGKDD international
Vol. 22. No. 3. 2011. conference on Knowledge discovery and data mining. ACM, 2012.
[20] Lim, Ee-Peng, et al. "Detecting product review spammers using rating [35] McAuley, Julian, Rahul Pandey, and Jure Leskovec. "Inferring networks
behaviors." Proceedings of the 19th ACM international conference on of substitutable and complementary products." Proceedings of the 21th
Information and knowledge management. ACM, 2010. ACM SIGKDD International Conference on Knowledge Discovery and
[21] Liu, Bing, et al. "Building text classifiers using positive and unlabeled Data Mining. ACM, 2015.
examples." Data Mining, 2003. ICDM 2003. Third IEEE International
Conference on. IEEE, 2003.

You might also like