Professional Documents
Culture Documents
Guided by,
Dr. S. Valli,
Professor,
College of Engineering, Guindy,
Anna University, Chennai.
Signature of Guide
Problem Statement
In today's world, where we have lots of information and news shared on social
websites, there's a big need for a system that can reliably check if news articles and
their sources are trustworthy. The problem at hand is to develop an effective
methodology to evaluate the trustworthiness and reliability of news content shared on
a social website by consider various factors, including the behaviour of users,
temporal patterns, and the influence of potential spam or malicious entities.
Phase 1:
On social websites, some users act dishonestly. They might pretend to be
someone else, or promote annoying stuff. To make the website safer and more reliable,
we need a way to spot these dishonest users and separate them from the honest ones.
These malicious actors engage in various deceptive activities, including spreading
false information, impersonating legitimate users, and promoting spammy content. To
mitigate the adverse effects of spam users and ensure the credibility of user
interactions, there is a need for an effective system to identify and categorize these
spam accounts reliably.
Phase 2:
In the world of social media, we see a lot of news and information. But not all
of it is trustworthy. Some news might be true, while others are not. The challenge is to
develop a systematic approach to assess the credibility of news content on a social
website. Using the insights from Phase 1, the goal is to provide users with accurate
and trustworthy information about the reliability of news articles and their respective
sources, thus enhancing the overall quality of user engagement with news content.
Related Works
Lingam [1] have proposed adaptive Deep Q-Learning for detecting Social Spam Users
who post spam content manipulate their behaviour by pretending as a normal user, and
perform phishing attacks. Experimentation was conducted using Social Honeypot
dataset. Precision, True Positive Rate, False Positive Rate, F-measure are the
performance metrics used for the project.
Josephine [2] have proposed Hybrid Particle Swarm Optimization with Deep Learning
Driven Sarcasm Detection (HPSO-DLSD) technique to identity sarcasm, a positive
sentence with negative meaning, on social platform which was considered to be
challenging problem in Natural Language processing domain. Experimentation was
conducted on Twitter and headline datasets. Recall, Precision, F-score, Accuracy are
the performance metrics used for the project.
Yang [3] proposes the idea of ranking meta path between 2 different users in a social
network by considering the direct influence, indirect influence and global influence of
user on others. It also assigns weightage to the speed range of influence parameter.
Experimentation was conducted on 4-area and Yelp dataset. Degree Centrality,
PageRank are the performance metrics used for the project.
Masood [4] proposes the user model evaluation with twitter accounts where the tweets
and features of various users are extracted and analysed. The classification process is
categorized as Fake content-based spammer detection, URL based spammer detection,
Spam in trending topics and Fake user identification using statistical data.
Experimentation was conducted on CREDBANK and PHEME dataset.
Wang [5] proposes a novel approach SentiDiff to merge textual information with
sentiment diffusion data within a supervised learning framework. The exploration of
sentiment diffusion was particularly focused on sentiment reversals where tweets and
their retweets exhibit differing sentiment polarities. Experimentation was conducted
on dataset obtained from Beijing Intelligent Starshine Information Technology
Corporation, a leading big data collection and mining service provider in China. Area
Under the Precision-Recall Curve (PR-AUC) was the performance metrics used for
the project.
Overall Architecture