Professional Documents
Culture Documents
Arunav Saikia
M.S. Data Science
Indiana University - Bloomington
arsaikia@iu.edu
Table 1: Four tweets from the OLID dataset, with their labels for each level and final categorization.
embeddings to detect abusive user comments on entities can take proper well informed mitigation
Wikipedia. steps. Our proposed method is a hierarchical ap-
proach with three sub-tasks:
In (15) Singh et al use CNNs and LSTMs to detect
aggression in social media text involving multilin- • Sub-task A: Offensive language identification
gual speakers. Detecting and filtering such content as offensive (OFF) or not offensive (NOT)
on social media is a challenging task because most
• Sub-task B: Automatic categorization of of-
multilingual populations use ‘code-mixing’ to ex-
fense types as either targeted insult (TIN) or
press their opinions. Code mixing refers to the
untargeted (UNT)
practice of using linguistic units from multiple lan-
guages in a single sentence. The authors in (15; 16) • Sub-task C: Offense target identification as in-
focus their work on detecting aggressive markers dividual (IND), group (GRP) or other (OTH)
on code mixed Hindi-English text generated on
Our end-to-end hierarchical framework allows us
Facebook and Twitter by Indians.
to use the combination of labels in each sub-task
All of the above approaches solely rely on linguis- to identify specific types of offensive language
tics and NLP to detect abusive content. But re- like profane or hate speech or cyberbullying etc
cent advances in graph mining and graph repre- as shown in Table 1
sentation learning has led to social network anal-
yses approach to tackle this problem. Emphasiz- 3 DATA
ing the principle of homophily, Mishra at el (17) The data used in this work was released as part of
mentioned that abusive content comes from users SemEval 2019 Task 6: Identifying and Categorizing
who share a set of common stereotypes and form Offensive Language in Social Media (OffensEval)
communities around them. The authors proposed (22). Offensive Language Identification Dataset
a representation learning approach to create ‘pro- (OLID) dataset, is a large collection of English
files’ of tweeter users. Specifically, the authors tweets annotated using a hierarchical three-layer an-
leveraged a node embedding framework called notation model to distinguish between whether the
node2vec (18) to learn node representations from language is offensive or not (A), its type (B), and
a community graph where nodes represent tweeter its target (C). It contains 14,100 annotated tweets
users and edges represent follower-following rela- divided into a training partition of 13,240 tweets
tionship. The authors used these profile representa- and a testing partition of 860 tweets. Each level is
tions along with text-based features from (20) and described in more detail in the following subsec-
a GBDT classifier to detect racist and sexist tweets. tions.
Most of these work do not distinguish between the
3.1 Level A:
different types of offensive language that exists
in the social media platforms. They either 1) use • Not Offensive (NOT): Posts that do not con-
terms like ‘cyberbullying’, ‘abusive’, ‘hate speech’ tain offense or profanity;
interchangeably when there exist clear definition
• Offensive (OFF): Posts containing any form
(21) for each form of offensive language based on
of non-acceptable language (profanity) or a
the content and the target or 2) work on specific
targeted offense, which can be veiled or di-
forms of offensive language like racist vs sexist
rect. This includes insults, threats, and posts
(7) while ignoring others. Identification and proper
containing profane language or swear words
categorization of all forms of offensive language is
important so that the organizations and government
A B C Train Test Total
OFF TIN IND 2,407 100 2,507
OFF TIN OTH 395 35 430
OFF TIN GRP 1,074 78 1,152
OFF UNT - 524 27 551
NOT - - 8,840 620 9,460
All 13,240 860 14,100
3.3 Level C:
• Individual (IND): Posts targeting an individ-
ual. This can be a famous person, a named Figure 2: Histogram of # words in a tweet from sub-task A
individual or an unnamed participant in the
conversation. Insults and threats targeted at
4 PROPOSED METHOD
individuals are often defined as cyberbulling.
5 RESULTS
• Group (GRP): Posts targeting a group of peo-
ple considered as a unity due to the same eth- 6 CONCLUSION
nicity, gender or sexual orientation, political
7 SOURCE CODE
affiliation, religious belief, or other common
characteristic. Many of the insults and threats Source code for all experiments and results can
targeted at a group correspond to what is com- be found here https://github.com/arunavsk/
monly understood as hate speech. OffenseEval2019.