Professional Documents
Culture Documents
SUBMITTED BY
ABEBE WALDESANBET GINA
ADVISOR
SUKHADAVE PRAMOD (PhD.)
AMBO
April. 2021
i|Page
Declaration
I, the undersigned, declare that the thesis comprises my own work. In compliance with
internationally accepted practices, I have dually acknowledged and refereed all materials
used in this work. I understand that non-adherence to the principles of academic honesty and
integrity, misrepresentation/ fabrication of any idea/data/fact/source will constitute sufficient
ground for disciplinary action by the university and can also evoke penal action from the
sources which have not been properly cited or acknowledged.
________________________________
________________________
Signature.
___________________
Date.
ii | P a g e
AMBO UNIVERSITY
SCHOOL OF GRADUATE STUDIES
CERTIFICATION SHEET
A thesis research advisor, I hereby certify that I have read and evaluated this thesis prepared under
my guidance by _______________________entitled “FAKE NEWS DETECTION ON
SOCIAL MEDIA BY USING DEEP LEARNING FOR AFAAN OROMO LANGUAGE”.I
recommend that it be submitted as fulfilling the thesis requirement.
iii | P a g e
Acknowledgment
First and foremost extraordinary thanks go for my Almighty God (Waaqa Gurraacha).
I would like to express my gratitude and heartfelt thanks to my advisor, Dr. Promod for his keen
insight, guidance, and unreserved advising. I am really grateful for his constructive comments and
critical readings of the study.
I am very grateful to the management and staff of Ambo University, especially to all Computer
Science staffs for constant support and providing appropriate professional comment.
I am immensely indebted to my beloved family especially, Ato. Ashenafi Tadessa I would like to
thanks you for you endeavor support and guidance throughout my life in any manner I am really
proud to have you.
The last but not the least, my sisters Tajitu W/sanbet you takes the role of mothers and you have
special place in my heart for giving me unconditional care, love, time, patient and support
throughout my life and thank you more (Galatoomi!)
My special thanks also goes to all my friend, family, that I did not mention your name and all
classmate for encouragement, support and the good friendship we shared during class.
iv | P a g e
Abbreviations
RNN- Recurrent Neural Networks
v|Page
Contents
Declaration .................................................................................................................................................... ii
Acknowledgment ......................................................................................................................................... iv
Abbreviations ................................................................................................................................................ v
Abstract ........................................................................................................................................................ xi
CHAPTER 1 ................................................................................................................................................ 1
1. INTRODUCTION............................................................................................................................... 1
1.1 Background Of The Study ............................................................................................................ 1
CHAPTER 2 ................................................................................................................................................ 9
2. REVIEW ON FAKE NEWS DETECTION ..................................................................................... 9
2.1 Introduction ................................................................................................................................... 9
2.3 Basic concept of fake news detection in deep learning models .................................................. 10
vi | P a g e
2.4 Recurrent neural networks (RNN) .................................................................................................. 15
2.5 Long short-term memory networks (LSTMS) ................................................................................ 17
2.6 Bi-directional long short-term memory. ......................................................................................... 18
2.7 Related works on fake news detection ........................................................................................ 19
CHAPTER 4 .............................................................................................................................................. 40
4 EXPERIMENT AND RESULTS ..................................................................................................... 40
4.1 Tools used ................................................................................................................................... 40
vii | P a g e
4.6 Experimental evaluation and results ........................................................................................... 46
CHAPTER 5 .............................................................................................................................................. 51
5. CONCLUSION AND RECOMMENDATION .............................................................................. 51
5.1 Conclusion .................................................................................................................................. 51
References .................................................................................................................................................. 53
Appendix ..................................................................................................................................................... 59
viii | P a g e
List of Tables
Table 3-1 Examples of statements and side information in the dataset. ....................................... 30
Table 4-1 Dataset specification. .................................................................................................... 40
Table 4-2 Specification of News Dataset 1................................................................................... 41
Table 4-3 Specification of News Dataset 2................................................................................... 41
Table 4-4 Concatenated dataset specification. .............................................................................. 42
Table 4-5 Bi-LSTM Confusion matrix model. ............................................................................. 47
Table 4-6 Experimental results ..................................................................................................... 48
Table 4-7 LSTM model details. .................................................................................................... 48
ix | P a g e
List of figures
Figure 1-1 SWOT Analysis ............................................................................................................ 8
Figure 2-1Biological Neural Network. ......................................................................................... 11
Figure 2-2 Simple Artificial Neuron. ............................................................................................ 11
Figure 2-3Neuron model with logistic activation function. .......................................................... 13
Figure 2-4 Neural networks with 2 hidden layers for 3-label classification. ................................ 15
Figure 3-1Octoparse implementation. .......................................................................................... 28
Figure 3-2 Partial sequence diagram for implemented scraper. ................................................... 29
Figure 3-3The basic seq2seq model. ............................................................................................. 33
Figure 3-5 Word cloud for text that is real. .................................................................................. 35
Figure 3-6 word cloud for text that is real. ................................................................................... 35
Figure 3-7 Architecture of Fake news detection based on Bi-directional LSTM-recurrent neural
network. ........................................................................................................................................ 36
Figure 3-8 Architecture of Fake news Detector using Flask. ....................................................... 39
Figure 3-9 Serialization and de-serialization. ............................................................................... 38
Figure 4-1 Keras summary of RNN model. .................................................................................. 43
Figure 4-2 General-architecture-of-Bi-directional-LSTM-RNN. ................................................. 44
Figure 4-3 Number of epochs applied for training. ...................................................................... 45
Figure 4-4 Bi-LSTM Classification report. .................................................................................. 48
Figure 4-5 System Interface for Afaan Oromo Fake News Detection. ......................................... 50
x|Page
Abstract
In recent years due to the booming development on internet, social media facilitates the creation
and sharing of information that uses computer-mediated technologies. This social media
development has changed the way groups of people communicate and interact. Now a day,
majority of people search and consume news from social media rather than traditional news
organizations these days. At one side, where social media have become a powerful source of
information and bringing people together, but identifying the inaccurate news is a difficult
problem. Without the concern about the credibility of the information, the unverified or fake news
is spread in social networks and reach thousands of users and get published, which can lead the
followers to negative effects or even manipulation of public or private events. One of the
distinctive challenges of detecting fake news on social media is how to identify fake news about
recent events. Most Previous works proposed appropriate machine learning models of fake news
detection and classification for English text. However, Afaan Oromo language difficulty in
semantic analysis due to morpheme ambiguity and lack of adequate datasets did not get more
attention. As a result, Afaan Oromo fake news detection is essential to maintain robust online
media and social network. We worked to resolve Social media fake news for Afaan Oromoo
language by implementing deep learning models and classifying them into the pre-defined fine-
grained categories. At first, we develop the paper presents a fake news detection models based on
Bi-directional LSTM recurrent neural network and representations obtained from these models are
fed into a Multi-layer Perceptron Model (MLP) for the final classification. The model is able to
give predictions on a benchmark dataset shows a results with an F1 score of 90%, which
outperforms the current state of the art.
The models were trained and evaluated on the Afaan Oromo language Fake News dataset Scrapped
from Twitter and Facebook. Finally, python spyder IDE was used for the web-based prototype
development of the trained Bi-LSTM models.
Key Words: Fake News detection; Deep learning; Bi-directional LSTM; Afaan Oromoo
xi | P a g e
CHAPTER 1
1. INTRODUCTION
1.1 Background of the Study
In the earliest times, long before the advent of computers and the web, fake news (also known as
deceptive news) were communicated through the oral tradition it may related event or issues in
public or private concern in the form of rumors (face to face), either to innocently talk about other
people lives, or to intentionally harm the reputation of other people or rival companies. Now a day
people need low cost and easy access to the information and dissemination quickly to push people
to search for news and know what is happening at the beginning of events. In recent years social
media play a crucial role over the traditional news transmission means and online content has been
playing a vital role in convincing user decisions and opinions(Ahmed et al., 2017) (Gereme & Zhu,
2019). Due to the booming of social networks fake news for commercial, politics and personal
interest has been widespread in the online world, users can get infected by these social media fake
news easily, which has brought about marvelous effects on the offline society already(J. Zhang et
al., 2020).
Now a day, due to different politics and economic events the huge amounts of information are
generated on the social networking with various social media’s formats(Conroy et al., 2015).
However, when some event has occurred, many people discuss it on the web through different
social media. Consequently it is possible that some fake news or misinformation are generated and
propagate in the chain throughout the social media, unintentional (or not) to lead other users with
intention of deception, misleading, grabbing attention or even financial and political gain(LIAO
& LIN, 2018)(Cardoso Durier da Silva et al., 2019).
Fake news is a sensitive message with its contents that claim people to believe with falsifications
and were received they will rapidly dispersed in chain of today’s digital world to other peoples.
The dissemination of fake news through internet has made the confusion of truth with taking
advantage of social media content to mislead readers and get published, which can lead to negative
effects or even manipulation of public events. Reports indicate that the human ability to detect
deception without special assistance is only 54% (Girgis et al., 2019)(Conroy et al., 2015).
1|Page
Fake News detection is a challenging task to accomplish, as it requires models to summarize the
news text and compare it to the actual news in order to classify it as fake or real news. However
technologies such as Artificial Intelligence (AI) and Natural Language Processing (NLP) tools
offer great potential for us to build systems which could automatically detect and classify fake
news. Moreover, the task of comparing proposed news with the original news itself is a daunting
task as its highly subjective and opinionated(Thota et al., 2018a).
There are different approach and way to detect fake news, stance detection is one of the way and
which will be the focus of our study. Stance Detection is the process of detecting the relationship
between two pieces of text which is fake and real. In this study, we explore ways to predict the
stance by calculate the semantic similarity between two piece of texts by using a deep learning
models.
Through experimental procedures, we used pre-trained model which can detect fake news by
accurately predicting stance between news articles. We also studied how different hyper
parameters affect the model performance and summarized the details for future work.
2|Page
are a number of media agencies produce an Afaan Oromo news articles in social media format.
Some of such are: VOA Afaan Oromo, BBC Afaan Oromo, Oromia communication bureau,
Oromia Broad cast service, etc.
Millions of news articles are being circulated every day on social media in Afaan Oromo text. As
a result of this, how one can trust which is real or fake. (Bahad et al., 2020)The news disseminated
on social media platforms may be of low quality carrying misleading information and comes at
the cost of dubious trustworthiness and significant risk of exposure to fake news and Afaan Oromo
text readers are not exceptional to suffer from this problem. It’s difficult to accurately distinguish
true from false information by just looking at these short pieces of information.
As a consequence of the above, fake news have become one of the major concerns because it’s
potential danger to modern society and there is no suitable hand engineering features model created
in the area which achieves the states-of-the arts results to identify fakeness of such statements.
The results of the research will contribute towards existing research fake news detection using
machine learning and attempt to answer the research question.
The specific objectives of the study are try to achieve the following:-
To review related research works in the area of fake news detection in different approach.
3|Page
To introduces the topic of fake news and deep learning algorithms that would be effective
in classifying fake news.
To scrap news documents from social media for training and testing model.
To develop a methodology to achieve the objective of the research and evaluate the result
of pre-trained model by using the evaluation metrics.
To develop System prototype for Afaan Oromo language fake news detection.
To presents a possible solution and draw conclusions based on experimental result as well
as lays some ground work in further study on this area.
1.5 Motivation
In the case of our country Ethiopia, fake news has the potential for extremely negative impacts on
individuals and society that led to a lot of problem. The aforementioned facts and figures show
that Afaan Oromo Language is the widely spoken language in the horn of Africa. With the fact
this language is used in schools, offices, and social media, there is the huge amount of data
available that encourage study related to deep learning tasks associated with language. The
motivation for research on this topic was that this is a relatively new area of research with many
opinions but not many concrete solutions.
The development of deep learning application for this language required to cope up with the
current technology to bring awareness, propose a solution, and work towards minimizing the
effects of fake news.
4|Page
1.7 Scope and limitation of the study.
This thesis focuses on fake news detection for Afaan Oromo news articles supervised
learning. Therefore, the experimentation has dealt with Afaan Oromo news texts only. On
the other hand, the absence of standard text corpus for Afaan Oromo language was a
limitation and the amount of data prepared for this study is relatively small and require
further enhancement of the size for further experimentation and evaluation.
5|Page
In this study we construct Afaan Oromo fake and real news datasets from Twitter and Facebook,
respectively and which consists 4500 real news and 2500 fake news. We name this data set as
ORORF2020 in this study. The amount of dataset prepared for the study is relatively larger than
(Zaman et al., 2020)news data set prepared for Indonesian language. However it’s smaller than
the data set prepared for English language and it requires more enhancement for further
experiment.
1.8.3 Preprocessing
Once data set is prepared, real-world data tend to be incomplete, noisy and inconsistent. This can
lead to a reduced quality of collected data, and, further to a low quality of models built on such
data. In order to address these issues, the data requires special preprocesses to implement deep
learning algorithm on them. We mainly focused on two issues: - Firstly the data must be organized
in proper form, for deep learning algorithms, and secondly, the data sets used must lead to the best
performance and quality for the models. To reduce the size of actual data, generic refinements like: -
stop-word removal, tokenization, a lower casing, and stemming were applying to removing the
irrelevant information that exists in the ORORF2020 data. We also provide insights into different
word vectors representations we used as part of our analysis.
6|Page
1.8.4 SWOT analysis
In the research of fake news detection for Afaan Oromoo language deep learning approach ,SWOT
Analysis is a useful technique for understanding the Strengths and Weaknesses of the work, in
addition to identifying both the Opportunities open to our work and the threats for the future.
We have used Bi-LSTM as a part of recurrent neural network it has designed for long term
dependencies, therefore the idea which makes it different and unique from other neural network is
that it is able to remember information for a long span of time without learning, again and again,
making this whole process simpler and faster, it includes an inbuilt memory for storing
information. There is no special training step or units added, the idea is just to read a sentence
forward and backward to capture more information and proving that Bi-LSTM representations are
more robust than representation learned by other models. Despite recent concerns and
controversies, YouTube is the leading social network of the world next to Facebook. However we
are limited to Afaan Oromoo news texts and as well as, the news texts we collect is not sufficient
enough when compared to the research for English language. Table 1.1 shows summary of SWOT
analysis of our work.
Strength Weakness
We are prepared around 7000 Afaan Our data set is not sufficient enough.
Oromoo news text and trained the Because, the model requires large
model. The implementation of model is amount of data for training.
simple and effective We did not consider to unlabeled data.
Rather than spending hours and human
powers to check the validity of news,
configure deep learning model to
manage the problem of fake news in
Afaan Oromoo News text.
7|Page
Opportunity Threats
Fake news detection is the hottest As opportunity when combine with
research area in deep learning the new technology and automate
approach. So, combine with newer complexity of algorithm may raise
forms of technology like IoT (Internet when it comes to implementation
of Things) allow us automate and
chances to make greater profits in the
environment
It can be applied to unlabeled data
Figure 1-1 SWOT Analysis
This thesis report is organized into five chapters. The first chapter talks about the motivation
behind conducting the research and discusses: background of the study, statement of the problem,
the objectives, methodology and scope of the study.
The remaining chapters of the paper are structured as follows:
Chapter 2: will give an overview of the literature underlying this research, for fake news detection
in the area and summarize the main approaches proposed to address this problem. Chapter 3: will
describe in detail materials and methods used to detect fake news.
Chapter 4: will present and discuss in detail the findings and results of the various experiments
carried to evaluate the proposed detection methods and the corresponding datasets used. Chapter
5: will make concluding remarks, by discussing the overall results of the research in the context of
the related work. In addition, it will suggest possible improvements and recommendation for future
works.
8|Page
CHAPTER 2
According to (Stahl, 2019) fake news contributor has classified into social bots, trolls and cyborg
user. If a social media account is being controlled by a computer algorithm, then it is referred to
as a social bot. However, fake humans are not the only contributors to the dissemination of false
information; real humans are very much active in the domain of fake news. As implied, trolls are
real humans who “aim to disrupt online communities” in hopes of provoking social media users
into an emotional response. While contributors of fake news can be either real or fake, what
happens when it’s a blend of both? Cyborg users are a combination of automated activities with
9|Page
human input. The accounts are typically registered by real humans as a cover, but use programs to
perform activities in social media. Acknowledging the impact of fake news, researchers are trying
different methodologies to find a quick and automatic solution to detect fake news in the recent
past years(Brien et al., 2018).
Neural networks are inspired by the way the human brain works. A human brain can process
huge amounts of information using data sent by human senses (especially vision). The
processing is done by neurons, which work on electrical signals passing through them and
applying flip-flop logic, like opening and closing of the gates for signal to transmit through(Kong
et al., 2020).
In biological setting (Figure 2.1)(Trung Tin, 2018), one neuron receives a signal from its tree of
dendrites, or dendritic tree, and if the signal is strong enough, it will pass through an axon and link
to a dendrite from another neuron. Two neurons are actually separate from each other by synaptic
gaps and only become connected when the link of an axon from one neuron and dendrite from the
others are stimulated.
10 | P a g e
Figure 2-1Biological Neural Network.
11 | P a g e
& Song, 2020). In other words, it reflects the influence of synaptic connection to the decision
whether or not the axon is stimulated. Similarly, we also have x2, x3,……..., xn that get multiplied
by W2, W3,...,Wn respectively. All of the products are then summed into one unit to depict
collective influence of those inputs. But whether the input is strong enough to make the neuron
fired? To model this, we take summation of all input neurons and put the result through an
activation function. If the of output of the activation function greater than 0, the axon is stimulated.
Figure 2.2 describes a neuron network with logistic activation function. In this case, activation
hW(x) is computed as:
1 + e−z (2.2)
In the same fashion, multiple connections are modeled by multiple layers with different sets of
weights. Suppose that we have a neural network with 3 layers as described in (Kresnakova et al.,
2019) Figure 2.2, activations of the hidden layer (layer 2) are computed as:
12 | P a g e
Figure 2-3Neuron model with logistic activation function.
Note that biases x0 and a(2)0 are omitted in this figure.
a(2)2 = g(W(1)20 x0 +W(1)21 x1 +W(1)22 x2 +W(1)23 x3) (2.5)
a(2)3 = g(W(1)30 x 0 +W(1)31 x 1 +W(1)32 x 2 +W(1)33 x 3) (2.6)
For the literature focused based on deep learning approach equation (2.3), (2.4), (2.5) are written
in matrix form.
Firstly, weight matrix representing the connection between layers 1 and layer 2 is written as:
13 | P a g e
Then,
Z (2) =
W (1) x (2.8)
a 02
a 12
a 2= a22 = g( z(2) ) (2.9)
a 32
Finally,
Where as
hW(x) =Output, α 3 = function, and g(z(3) are net input calculated
14 | P a g e
Figure 2-4 Neural networks with 2 hidden layers for 3-label classification.
15 | P a g e
Figure 2-5 an unrolled recurrent neural network.
Given input xt and hidden state of previous step ht-1, new hidden state and output at time step t is
computed as:
ht = σh(Whxt + Uhht−1 + bh) (2.13)
yt = σy(Wy ht + by) (2.14)
Where:
xt is input vector at time step, ht is hidden layer vector, yt is output
vector at time step t.
W,U,b are parameter matrices and vector.
σh, σy are activation functions.
Recurrent neural network is particularly designed to deal with sequential data where
inputs are not fed into the networks all at once, but are broken down into small pieces
which are later passed into the network cell one after another. Despite being designed
to deal mimic and work on sequence nature of some kinds of data, it is proved that
RNNs have limitations in capturing long dependencies(Kresnakova et al., 2019). As a
result, Long Short-Term Memory Network, a modified version of RNN with gating
mechanisms, is devised to get over the limitation of vanishing gradient Problem in the
layers of deep neural network.
16 | P a g e
Figure 2-6 internal structure of Long Short-Term Memory Networks.
17 | P a g e
2.6 Bi-directional long short-term memory.
Long short-term memory (LSTM) is a structure that learns how much of the previous network
state to apply when input data is received. It preserve the error that can be back-propagated through
time and in lower layers of a deep network is a sequence processing model that consists of two
LSTMs: one taking the input in a forward direction, and the other in a backwards direction(Yulita
et al., 2017)(Bahad et al., 2020). Figure 2.8 shows, a Bi-Directional LSTM network steps through
the input sequence in both directions at the same time. It resolves the long-term dependency
problem of conventional recurrent neural network (RNN) using both the hidden state and the cell
state, which is a memory for storing past input information and the gates that are used to regulate
the ability to remove or add information to the cell state. The multiplicative gates and memory are
defined for time t(Kong et al., 2020)
18 | P a g e
ht= ot * tanh( Ct ) 2.19
Where, σ (⋅) is the sigmoid function and ft, it, ot, Ct, and ht, ft, it, ot, Ct, and ht are the vectors of
the forget gate, input gate, output gate, memory cell, and hidden state, respectively. All of the
vectors are the same size.
Moreover, Wf, Wi, Wo, and Wc Wf, Wi, Wo, and Wc denote the weight matrices of each gates
and bf, bi, bo, and bc, bf, bi, bo, and bc denote the bias vectors of each gates. Another shortcoming
of conventional RNN is that they are only able to make use of previous context(Kong et al., 2020).
To resolve this, bidirectional-RNN (Bi-RNN) stacks two RNN layers. If the existing RNN is the
forward RNN that only forwards previous information, Bi-RNN stacks backward RNN that can
receive subsequent information, as shown in Fig. 8. Combing Bi-RNN with LSTM gives
Bidirectional- LSTM (Bi-LSTM), which can handle long-range context in both input
directions(Kong et al., 2020).
19 | P a g e
engineering. (Ma et al., 2016)Modeled streams of tweets as sequential data, then used Recurrent
Neural Network (RNN) for predicting weather the streams were fake or not. This approach was
proved to yield better results than previous feature-based learning and effective at early rumor
detection. The researcher outperforms existing model and able to achieve better accuracy.
21 | P a g e
2.9 Related work on local language
Research on fake news detection in Ethiopian local language is still at an early stage. (Gurmessa,
2020b) describe the application of natural language processing techniques with multinomial naïve
Bayes for the detection of fake news on 752 Afaan Oromoo news text. They were used facebook
as the source of news article and apply term frequency-inverse document frequency (TF-IDF) of
unigram and bi-grams. With the best f1 score they were achieved good results based on their data
size. However, as well as the accuracy obtained is good the source of data set site’s information
are not 100% credible and the most accuracy could have been attained by considering non-credible
news into account. In order to building block for Fake news detection, (Gurmessa, 2020a) were
develop Content-Based Afaan Oromo fake news detection system using a machine learning
approach with a passive-aggressive classification algorithm. They were used the data set collected
manually from facebook pages and labeled as fake and real. Since the data set is critical issues in
Afaan Oromo language, the classification was tested on a small number of news dataset items and
increase the size of data to the news dataset test the consistency of the performance thereby
increasing the trust of users on the system.
22 | P a g e
hand, do not differ greatly from English letters, however there are few special combinations which
is called “Qubee Dachaa” such as “sh” and “ch” (same sound as English), “dh” in Afaan Oromo
is like an English "d" produced with the tongue curled back slightly and with the air drawn in, so
that a glottal stop is heard before the following vowel begins. Another combination is “ph” made
when with a smack of the lips toward the outside “ny” closely resembles the English sound of
“gn”. We commonly use these few special combination letters to form words. For example, dh
use in dhadhaa ‘butter’ , ch used in barbaachisaa ‘important’, sh used in shamarree ‘girl’, ph
used in buuphaa ‘egg’, and “ny” used in nyaata ‘food’ .
Afaan Oromo language has 36 letters (26 consonants and 10 vowels) called “Qubee”. Words in a
sentences are separated by white spaces the same way as it is used in English. Different Afaan
Oromo punctuation marks follow the same punctuation pattern used in English and other languages
that follow Latin writing system. For example, comma (,) is used to separate listing of ideas,
concepts, names, items, etc. and the full stop (.) in statement, the question mark (?) in interrogative
and the exclamation mark (!) in command and exclamatory sentences mark the end of a
sentence(Tesfaye, 2010a). In general, all letters in English language are also in Afan Oromo
language, except the way it is written.
In Afaan Oromoo languages Vowel can appear in initial, medial and final positions in a word in
Afaan Oromo language. A long vowel is interpreted as a single unit and occurs everywhere a short
vowel can occur.
The following examples show some of long vowels at word initial, medial and final positions.
Initial positions: eelee to mean ‘pan’, uumaa to mean ‘nature’,
Medial position: leexaa to mean ‘single’ keennaa to mean ‘gift’,
Final position: garaa to mean ‘belly’, daaraa to mean ‘ash’
The difference in length is contrastive, for example consider, ‘lafa’ in Afaan Oromoo which is to
mean 'land', and ‘laafaa’ in Afaan Oromoo which is to mean 'weak'. The difference between the
words ‘lafa’ and ‘laafaa’ is the length of vowel they have. Two vowels in succession indicate that
the vowel is long (called “Dheeraa” in Afaan Oromoo), while a single vowel in a word is short
(called “Gababaa” in Afaan Oromoo).
The difference in length is contrastive, for example consider, ‘lafa’ in Afaan Oromoo which is to
mean 'land', and ‘laafaa’ in Afaan Oromoo which is to mean 'weak'. The difference between the
words ‘lafa’ and ‘laafaa’ is the length of vowel they have. Two vowels in succession indicate that
23 | P a g e
the vowel is long (called “Dheeraa” in Afaan Oromoo), while a single vowel in a word is short
(called “Gababaa” in Afaan Oromoo). Table 2.1 Afaan Oromoo Vowels.
Front Central Back
High i, ii u, uu
Mid e, ee o, oo
Low a, aa
Table 2-1 Afaan Oromoo Vowels
Afaan Oromo vowels are pronounced in sharp and clear fashion which means each and every word
is pronounced strongly.
For example:
A: Fardda, haadha
E: Gannale, Waabee, Roobale, Colle
I: Arsii, Laali.
O: Oromo, Cilaalo, Haro, Caancco, Danbidoollo
U: Ulfaadhu, Arbba.
24 | P a g e
2.10.3 Afaan Oromo morphology
Every language has its own morphological structure that defines rules used for combining the
different components the language may have. The English language for instance is basically
different in its morphological structure from French, Arabic or Afaan Oromoo(Jimalo, P, et al.,
2017). There are a number of word formation processes in Afaan Oromoo. Affixation and
Compounding are among these word formation processes.
Affixation is generally described as the addition of affixes at the beginning, in between and/or at
the end of a root/stem depending on whether the affix is prefix, infix or suffix. Attaching one or
more prefixes and/or suffixes to a stem may form a word. The word durbumma ‘girlhood’ for
instance is formed from the stem durb- ‘girl’ and the suffix –umma.
Compounding is the joining together of two linguistic forms, which functions independently.
Examples compound nouns include; abbaa-buddenaa ‘step father’ from abba- ‘father’ and
buddena ‘food’. Like a number of other African and Ethiopian languages, Afaan Oromo has a
very complex and rich morphology. It has the basic features of agglutinative languages involving
very extensive inflectional and derivational morphological processes. In agglutinative languages
like Afaan Oromo, most of the grammatical information is conveyed through affixes (i.e. prefixes
and suffixes) attached to the root or stem of words. Obviously, these high inflectional forms and
extensive derivational features of the language are presenting various challenges for text
processing and information retrieval experiments in Afaan Oromo(Tune et al., 2008)(Tesfaye,
2011). Although, Afaan Oromo words have some prefixes and infixes, suffixes are the
predominant morphological features in the language. Almost all Oromo nouns in a given text have
persons, number, gender and possession makers which are concatenated and affixed to a stem or
singular noun form. In addition, Afaan Oromo noun plural markers/forms can have several
alternatives. For instance, in comparison to the English noun plural marker s (-es), there are more
than ten major and very common plural markers in Afaan Oromo including: -oota, -wwan, -lee, -
an, -een, -eeyyii, -oo, etc...).
As an example, the Afaan Oromo singular noun “mana” (house) can take the following different
plural forms: Manoota (mana + oota), manneen (mana + een). The construction and usages of
such alternative affixes and attachments are governed by the morphological and syntactic rules of
the language.
25 | P a g e
2.11 News writing structure on social media
Social media news, often referred to simply as social news, refers to a more modern tendency to
get news what is happening around us from social media rather than more traditional news sources.
it may involve current events, new initiatives, or other issues.
News writing structure or style is the way in which elements of the news are presented based on
relative importance, tone and intended audience. In addition, it is also concerned with the structure
of vocabulary and sentences(Tantiponganant & Laksitamas, 2014).
News writing attempts to answer all the basic questions about any particular event - who, what,
when, where and why (the Five W’s) and also often how - at the opening of the article. This form
of structure is sometimes called the "inverted pyramid", to refer to the decreasing importance of
information in subsequent paragraphs. The most important structural element of a story is the lead
which is contained in the story’s first sentence. The lead is usually the first sentence, or in some
cases the first two sentences, and is ideally 20-25 words in length (Tantiponganant & Laksitamas,
2014).
26 | P a g e
CHAPTER 3
27 | P a g e
Figure 3-1Octoparse implementation.
In this research work, we are scraping the official Twitter account of VOA Afaan Oromo and BBC
Afaan Oromo. As you can see on Fig 3.1, the website is loaded in the built-in browser and usually,
twitter account websites have a next page button that allows Octoparse to click on and go to each
page to grab more information from each page. Octoparse scraping the selected information by
scroll down the page and extract the tweets.
Figure 3.2 shows the sequence of steps implemented scraper in which twitter applies infinite
scrolling technique that we need to extract the data shown on the screen.
28 | P a g e
Figure 3-2 Partial sequence diagram for implemented scraper.
First, we build an extraction loop to extract the title and body of tweet fields one by one into
separate columns instead of just one, so we need to modify the extraction settings to select our
target extraction data manually. We have built a pagination loop earlier, but we made modification
set up the AJAX time and both the scroll repeats and the waiting time on the workflow setting.
Finally when the extraction is completed, we export the VOA and BBC Afaan Oromo tweets into
CSV file.
29 | P a g e
In this study, the size of the dataset to be extracted is 7000*3. It means that there are 7000 rows
along with 3 columns. The name of the columns are “Headline”, “Body” and “Label”. It is being
seen from the dataset approximate 3000 fake news articles and 4000 real news articles.
30 | P a g e
We have deal with two-label setting (i.e. True, fake) although the statements in the dataset are,
news articles considered to be on different topics of the community, social, economic,
technological and political issues so that they are a potential source for collecting balanced corpus
for the task of fake news detection for Afaan Oromo Language.
We summarize the following core techniques in terms of representing texts for the most
fundamental part of deep learning.
Word-level tokenizing functions for English texts from natural language toolkit (NLTK) libraries
it might not be a suitable method if we tokenize Afaan Oromo texts because, Afaan Oromo words
consists of various pre and postpositions. So we need to use Morpheme as token in Afaan Oromo
words because it is the smallest unit with meaning.
3.2.2 Stemming
Stemming is a technique for the reduction of words to their stems or root variant(Korenius et al.,
2004). Documents were contained several existences of words like “barattoota”, ”barattootni”,
“baratichi”, “baratichatu” , ”barattu‘, “barataa” Different words share the same word stem (i.e.
“barat”) and a module which was designed for this particular purpose used to convert those
different representations to their stems.
Doing this reduce computing time and space as different forms of words are stemmed to a common
word. In this thesis work a module which was developed to steam Afaan Oromo words designed
to remove suffixes from each words are: - Among those suffixes and prefixes some of them are
listed below.
Set of suffixes: - “oota”, ”icha”,‘tu‘, ”ichan”, ”ootni”, ”aa”, ”ichatu”, ”een”,…etc.
31 | P a g e
Set of Prefixes:- ”al”, ”hin” , ”ni”, ”wal”……etc.
32 | P a g e
Figure 4-3 the basic seq2seq model.
Fig. 4.3 shows a basic RNN seq2seq model with a bi-directional LSTM encoder and an LSTM
decoder. The bidirectional LSTM is considered since it usually gives better document
representations compared to a forward LSTM. The encoder reads a sequence of input tokens x and
turns them into a sequences of hidden states h = (h1, h2, h3,……….., hj ) with following updating
algorithm:
it = σ ( Wii Ext-1 + bii +Whiht-1 + bhi) (Input Gate)
ft = σ ( Wif Ext-1 + bif +Whfht-1 + bhf ) (Forget Gate)
ot =σ ( Wio Ext-1 + bio +Whoht-1 + bho) (Output Gate)
gt = tanh (WigExt-1 + big +Whght-1 + bhg)
ct = ftct-1 + itgt
ht = ot tanh(ct)
Where σ is the logistic sigmoid function, i, f, o, and c are the input gate, forget gate, output gate
and cell activation vectors, and vector b are learnable parameters, Ext denotes the word embedding
of token xt and ct represents the cell states.
33 | P a g e
For the bi-directional LSTM, the input sequence is encoded as → he and ←he, where the right and
left arrows denote the forward and backward temporal dependencies, respectively.
Superscript e is the shortcut notation used to indicate that it is for the encoder. During the decoding,
the decoder takes the encoded representations of the source article as the input and generates the
summary y. In a simple encoder-decoder model, encoded vectors are used to initialize hidden and
cell states of the LSTM decoder.
In an attention based encoder decoder architecture (shown in Fig. 3.1), the decoder not only takes
the encoded representations (i.e., final hidden and cell states) of the source article as input, but also
selectively focuses on parts of the article at each decoding step and this attention can be achieved
by alignment mechanism(Bahdanau et al., 2015).
34 | P a g e
Figure 3-4 Word cloud for text that is real.
35 | P a g e
3.4 SYSTEM ARCHITECTURE
Figure 3.1 is a diagrammatic representation of the classification process. It illustrates the steps that
were involved in this research from text-preprocess to final prediction of entire texts.
Figure 3-6 Architecture of Fake news detection based on Bi-directional LSTM-recurrent neural
network.
In order to create an artifact capable of detecting fake news, in the first step news text comes in
different format, from social media and undergoes set of preprocessing to eliminate unwanted
characteristics left by the data acquisition phase, and the cleaning phase more actions than non-
textual characters, such as fixing spelling and syntax errors, standardizing data sets, and correcting
mistakes such as empty fields, missing values in addition to stop word removal, Tokenization and
stemming.
36 | P a g e
Then, after preprocess and padded the sequence of words transform to numbers, because computers
cannot read words, and word embedding work so well as well as the semantics of the words are
captured. After the words are converted into word embedding’s, the words are fed into a neural
network. This neural network consists of various layers. The first layer is a convolutional layer. A
convolution is a filter that can extract features from the data and max pooling layer iterates over the
tensors and takes the highest value. After representation the data is classified as train and test. The
training is carried out on the news article corpus and test data is used to know the predicted label
of news article based on trained model, Bi-Directional LSTM network each embedding layer
corresponding to training data is inspected in both orders at the same time and once the classifier
was trained to separate targets from clutter, a threshold was applied to the output score to determine
the output.
In order to serialize the model using some particular format after training, and de serialize that
model in the production environment. Commonly, Python is the language for deep learning
modeling which have different serialization recommendations. In particular:-Sklearn recommends
using Joblib package and Pickle is used for serializing and de-serializing Python object structures.
Serialization refers to the process of converting an object in memory to a byte stream that can be
stored on disk. The trained model is saved to disk using pickling by which we need not train the
model every time we need it. We just need to de serialize it for predicting news input(Srivastava,
2020). Figure 3.8 shows, a way to write a python object on the disk that can be transferred
anywhere and later de-serialized (read) back by a python script.
37 | P a g e
Figure 3-8 Serialization and de-serialization.
The first step would be to load the saved pipeline and requested query, then computes the predicted
text news based on our model. For this de- serialized the pickled model receives the inputs and
uses the trained model to make the prediction and returns that prediction result to get the labels
which can be accessed through the API endpoint.
Finally, system prototype deployment of trained deep learning models are available to the end
users or systems. However, there is complexity in the deployment of deep learning models. Figure
3.9 shows deployment of our trained deep learning models into production using Flask API.
38 | P a g e
Figure 3-9 Architecture of Fake news Detector using Flask.
39 | P a g e
CHAPTER 4
We describe in this chapter the experiment evaluation of our approach and discuss obtained results.
For experiment of evaluation all cade are written in Python 3.7, using TensorFlow 2.3.1, Spyder
1.4.1 and NumPy. All experiments have been performed on a Core™ processor Intel® CPU i5-
3340M CPU @2.70 GHz with 4.00 GB RAM.
40 | P a g e
Table 4-2 Specification of News Dataset 1
41 | P a g e
Table 4-4 Concatenated dataset specification.
Data pre-processing is an important step here, as in most NLP applications(Z. Zhang & Luo, 201
9). The Title and Body of the article are concatenated as FakeReal, followed by removal of stop
words, tokenization and lemmatization of text. While tokenizing, a maximum of 150559 words is
considered. All textual sequences are then converted to numerical sequences and padded or trimm
ed to a maximum sequence length that we set.
42 | P a g e
from this GloVe embedding. Since the learned word weights in this model are not to be updated,
therefore the trainable attribute for this model is set to be false.
As embedding layer passes the deeper understanding of the input data to the next layers and learns
the dense vectors representation. In other tests later, it is initialized with pre-trained embedding.
43 | P a g e
The optimization function used was Adam and the loss was computed using binary cross-entropy.
The model was fit in 200 epochs with a batch size of 256 samples. Within the first few thousand
samples being run through the dataset the loss was already approaching 1, and by the end of the
first epoch it had hit 1. Here, the input sequence is a series of integers representing an index in the
word index dictionary. Each integer in the sequence corresponds to a word that had been on that
place before the encoding. During training we follow the following entire procedures:-
The encoder accept an input source sequence and computes the state values. Then, the encoder
passes final state vector as the initial state of the decoder. From the initial symbol, the decoder
sequentially predict the most probable target token that comes to the next. Figure 4. 2 shows overall
description of architecture, training a seq2seq network.
44 | P a g e
embedding vector in a target sequence and maximizes the probability of the next true target token
Y j+1 for the following time steps j = 1,…………,T2.
Note that the final output token is end of sequence. In summary, the model is trained to maximize
each target sequence [Y1, Y2,…………,YT2, end of sequences’ probability using both [X1
,X2,……….,XT1] and start of sequence, [Y1,Y2,…………,YT2 ] for the encoder and decoder input.
To alleviate overfitting and to escalation the generalization capacity of the neural network, the
model should be trained for an optimal number of epochs. Figure 4.3 .Shows rounds of
optimization that are applied during training.
45 | P a g e
sequence, the decoder works in a different ways from training. From the starting symbol start-of-
sequence (SOS) with the initial state from the encoder, it calculates the probabilities for all tokens
and finds the token that matches the highest probability. Then the decoder uses it as the input of
the next time step. It stops the prediction as the predicted token is end-of-sequence (EOS). To
evaluate how well the model performed on the predictions we have used confusion matrix to check
the number of correct and incorrect predictions.
46 | P a g e
In this experiment the Afaan Oromoo News text dataset were pre-processed and embedding with
creating a dense vector space for each word instead of traditional Bag of Words. We divided the
dataset into two sets: - 20% for testing set and 80% training set. While running both the cases for
number applied for each epochs, and a batch size of 128, accuracy and data loss changes are quite
visible, depicted in Figure 4.3 Number of epochs applied for training. An initial set of experiments,
were run using Bi-LSTM and the performance of the classifiers is also measured using F-measure
values, and outperform 90% of accuracy is obtained by using Bi-LSTM. When we feed
preprocessing data to an outstanding pre-trained model and of course, get output in probabilities,
better effectiveness and performance measurement of the model is good, compared to related work
(Chapter 2) and the Confusion matrix comes into the limelight. Precision and recall for each label
are also high (Table 4.2). More specifically, the following are observed:-
All examples labelled as “real” are classified correctly, with a class recall of 97% and a class
precision of 88% and all examples labelled as “fake” with class recall of 75% and a class precision
of 93%. Table 4.2 shows Confusion Matrix performance measurement for fake news detection in
deep learning model for Afaan Oromoo news text, with 4 different combinations of predicted and
actual values.
Predicted Fake Predicted Real Class precision
Actual Fake 352 72 93%
47 | P a g e
Figure 4-4 Bi-LSTM Classification report.
In the first experiment we used Vanilla RNN, then we used LSTM in the second experiment,
finally, in the last experiment we used the Bi-LSTM technique. The following table illustrates the
accuracy of each model.
Model Accuracy
LSTM 0.87
Bidirectional-LSTM 0.90
Vanilla-RNN 0.89
Table 4-6 Experimental results
As we can see from the results (Table 4.6), all models achieved very good performance. In the
model training, but the downside of RNN model suffers vanishing gradient problem due to the
deep network hierarchy. Bi-LSTM solves the vanishing gradient problem and the model has higher
prediction accuracy than the other model. Bi-LSTM model managed to detect all fake news articles
in the testing set, while still kept the FP ratio at a reasonable level. Table 4.2 depicts in more detail
the performance of the model using a confusion matrix.
It shows, that all positive examples (fake news articles) were correctly classified by the model,
while 352 of true news were predicted as fake ones.
48 | P a g e
4.7 Discussion
We examined the results of experiments, how the recurrent neural network handles sequential data
in detail as well as the main concept to build an RNN-based model from processing raw input
sentences to generating target sentences. Recurrent neural network (RNN) based architectures are
also proposed for Afaan Oromoo language fake news detection. RNNs process the word
embedding in the text sequentially, one word of token at a time, utilizing at each step the
information from the current word to update its hidden state which has aggregated information
about the previous sequence of words and the final hidden state is generally taken as the feature
representation extracted by the RNN for the given input sequence. However, the amount of training
required which is comparatively much more than other kinds of networks and RNNs suffer from
the problem of vanishing gradients, which hampers learning of data sequences and gradients carry
information used in the Recurrent Neural Network parameters update and when the gradient
becomes smaller, the parameter updates become insignificant which means no real learning is done
and to overcome the problem which alleviate some of the training difficulties in RNN we used
LSTM to its ability to effectively capture long range dependencies in the sequence of text, and
has been applied to fake news detection. The size of the data can affect the accuracy of the model
and as the size of the data increases, accuracy also increases as well as the number of training
increases. The shortcoming of the current model is that it requires more training data and training
time than the existing baselines. Further along a line as a future work, the inclusion of features like
source or the author of the article, user response, along with the model proposed, and increased
the volume of dataset can lead the way towards a state of- the-art solution to this potentially
hazardous “digital wildfire “now day. Finally based on the classification result Bi-LSTM was
chosen as the best model to determine the truth of Afaan Oromo news in social media and the
system prototype was developing and the user interface screenshot sample is shown in (Figure 12).
49 | P a g e
Figure 4-5 System Interface for Afaan Oromo Fake News Detection.
50 | P a g e
CHAPTER 5
51 | P a g e
1.2 Recommendation and future work
In this research, an attempt is made to design methodology for Afaan Oromo fake news texts
detection. Future work will take into account other traits of Afaan Oromoo fake news detection, in
hopes we think the most successful approach would be automatic fact checking model in Afan
Oromo language texts and other local languages, that is, compelling the model with some kind of
knowledge base, the purpose of the model would then be to extract information for the text and
verify the information in the database. The problem with this approach would be that the
knowledge base would need to be constantly and manually update to stay up to date in addition to
multi-class prediction. The following are some of the recommendations and future work for further
research and enhancement:-
The stop-word list used in this research is compiled during the data preparation and mostly
is news specific. The availability of standard stop-word list would definitely facilitate
researches in the area of prediction of fake news and other classification technique,
therefore a standard Afaan Oromo stop-word list should be developed.
This research considers only multi-label classification, it does not consider multi-class
classification, and therefore, research in this direction would also improve the classification
quality and furthermore this study focusing on text-based fake news. However, now a day
Social media posts with video have more views and fake information appears in a variety
of forms, including videos, audio, images, and text. For future work this gap must be
considered.
This study has been based on learning based approach. Future work could be make
knowledge based to be more confidence on the system.
As the precision, recall, F1 scores of all the models were almost the same, they were not
considered as the main evaluation metric. Future work could possibly make more use of
these metrics to make a more thorough comparison.
52 | P a g e
References
Ahmed, H., Traore, I., & Saad, S. (2017). Detection of Online Fake News Using N-Gram
Analysis and Machine Learning Techniques. Lecture Notes in Computer Science (Including
Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics),
10618 LNCS(October), 127–138. https://doi.org/10.1007/978-3-319-69155-8_9
Bahad, P., Saxena, P., & Kamal, R. (2020). ScienceDirect ScienceDirect Fake News Detection
using Bi-directional LSTM-Recurrent Neural Fake News Detection using Bi-directional
LSTM-Recurrent Neural Network Network. Procedia Computer Science, 165(2019), 74–
82. https://doi.org/10.1016/j.procs.2020.01.072
Bahdanau, D., Cho, K. H., & Bengio, Y. (2015). Neural machine translation by jointly learning
to align and translate. 3rd International Conference on Learning Representations, ICLR
2015 - Conference Track Proceedings, September.
Bajaj, S. (2017). “ The Pope Has a New Baby !” Fake News Detection Using Deep Learning. 1–
8.
Basaldella, M., Antolli, E., Serra, G., & Tasso, C. (2018). Bidirectional LSTM recurrent neural
network for keyphrase extraction. Communications in Computer and Information Science,
806(December), 180–187. https://doi.org/10.1007/978-3-319-73165-0_18
Biswas, E., Vijay-Shanker, K., & Pollock, L. (2019). Exploring word embedding techniques to
improve sentiment analysis of software engineering texts. IEEE International Working
Conference on Mining Software Repositories, 2019-May(May), 68–78.
https://doi.org/10.1109/MSR.2019.00020
Brien, N., Latessa, S., Evangelopoulos, G., & Boix, X. (2018). The Language of Fake News:
Opening the Black-Box of Deep Learning Based Detectors. 32nd Conference on Neural
Information Processing Systems, Nips, 1–5. https://dspace.mit.edu/handle/1721.1/120056
Cardoso Durier da Silva, F., Vieira, R., & Garcia, A. C. (2019). Can Machines Learn to Detect
Fake News? A Survey Focused on Social Media. Proceedings of the 52nd Hawaii
International Conference on System Sciences, 6, 2763–2770.
https://doi.org/10.24251/hicss.2019.332
53 | P a g e
Castillo, C., Mendoza, M., & Poblete, B. (2011). Information credibility on Twitter. Proceedings
of the 20th International Conference Companion on World Wide Web, WWW 2011, 675–
684. https://doi.org/10.1145/1963405.1963500
Chaudhry, A. K., Baker, D. & Thun-Hohenstein, P. (2017). Stance Detection for the Fake News
Challenge: Identifying Textual Relationships with Deep Neural Nets. Stanford, 1–10.
Conroy, N. J., Rubin, V. L., & Chen, Y. (2015). Automatic deception detection: Methods for
finding fake news. Proceedings of the Association for Information Science and Technology,
52(1), 1–4. https://doi.org/10.1002/pra2.2015.145052010082
Extension, A., & Sodo, W. (2018). RESEARCH ARTICLE THE IMPACTS OF FAKE NEWS ON
PEACE AND DEVELOPMENT IN THE WORLD : THE CASE STUDY OF ETHIOPIA * Dr
. Kiros Abeselom.
Fang, Y., Gao, J., Id, C. H., Peng, H., & Wu, R. (2019). Self Multi-Head Attention-based
Convolutional Neural Networks for fake news detection. 1–13.
https://doi.org/10.1371/journal.pone.0222713
Fikadu Dinsa, E., & Babu P, R. (2019). Application of Data Mining Classification Algorithms
for Afaan Oromo Media Text News Categorization. International Journal of Computer
Trends & Technology, 67(7), 73–79. https://doi.org/10.14445/22312803/ijctt-v67i7p112
Gereme, F. B., & Zhu, W. (2019). Early detection of fake news “before it flies high. ACM
International Conference Proceeding Series, 142–148.
https://doi.org/10.1145/3358528.3358567
Girgis, S., Amer, E., & Gadallah, M. (2019). Deep Learning Algorithms for Detecting Fake
News in Online Text. Proceedings - 2018 13th International Conference on Computer
Engineering and Systems, ICCES 2018, November, 93–97.
https://doi.org/10.1109/ICCES.2018.8639198
Gurmessa, D. K. (2020a). Afaan Oromo Fake News Detection Using Natural Language
Processing and Passive-Aggressive. December.
Gurmessa, D. K. (2020b). Afaan Oromo Text Content-Based Fake News Detection using
Multinomial Naive Bayes. 01, 26–37.
54 | P a g e
Hochreiter, S., & Schmidhuber, J. (1997). Long Short-Term Memory. Neural Computation, 9(8),
1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735
Jimalo, K. M., Babu P, R., & Assabie, Y. (2017). Afaan Oromo News Text Categorization using
Decision Tree Classifier and Support Vector Machine: A Machine Learning Approach.
International Journal of Computer Trends and Technology, 47(1), 29–41.
https://doi.org/10.14445/22312803/ijctt-v47p104
Jimalo, K. M., P, R. B., & Assabie, Y. (2017). Afaan Oromo News Text Categorization using
Decision Tree Classifier and Support Vector Machine : A Machine Learning Approach.
47(1), 29–41.
Kong, S. H., Tan, L. M., Gan, K. H., & Samsudin, N. H. (2020). Fake News Detection using
Deep Learning. ISCAIE 2020 - IEEE 10th Symposium on Computer Applications and
Industrial Electronics, 102–107. https://doi.org/10.1109/ISCAIE47305.2020.9108841
Korenius, T., Laurikkala, J., Järvelin, K., & Juhola, M. (2004). Stemming and lemmatization in
the clustering of finnish text documents. International Conference on Information and
Knowledge Management, Proceedings, January, 625–633.
https://doi.org/10.1145/1031171.1031285
Kresnakova, V. M., Sarnovsky, M., & Butka, P. (2019). Deep learning methods for Fake News
detection. IEEE Joint 19th International Symposium on Computational Intelligence and
Informatics and 7th International Conference on Recent Achievements in Mechatronics,
Automation, Computer Sciences and Robotics, CINTI-MACRo 2019 - Proceedings,
November, 143–148. https://doi.org/10.1109/CINTI-MACRo49179.2019.9105317
Kunapareddy, R., Madala, S., & Sodagudi, S. (2019). False content detection with deep learning
techniques. International Journal of Engineering and Advanced Technology, 8(5), 1579–
1584.
Kwon, S., Cha, M., Jung, K., Chen, W., & Wang, Y. (2013). Prominent features of rumor
propagation in online social media. Proceedings - IEEE International Conference on Data
Mining, ICDM, 1103–1108. https://doi.org/10.1109/ICDM.2013.61
Leea, H., & Song, J. (2020). Understanding recurrent neural network for texts using English-
55 | P a g e
Korean corpora. Communications for Statistical Applications and Methods, 27(3), 313–326.
https://doi.org/10.29220/CSAM.2020.27.3.313
Ma, J., Gao, W., Mitra, P., Kwon, S., Jansen, B. J., Wong, K. F., & Cha, M. (2016). Detecting
rumors from microblogs with recurrent neural networks. IJCAI International Joint
Conference on Artificial Intelligence, 2016-Janua, 3818–3824.
Monti, F., Frasca, F., Eynard, D., Mannion, D., & Bronstein, M. M. (2019). Fake News
Detection on Social Media using Geometric Deep Learning. 1–15.
http://arxiv.org/abs/1902.06673
Of, D., For, A. S., & Text, A. (2015). SCHOOL OF GRADUATE STUDIES SCHOOL OF
INFORMATION SCIENCE DEVELOPMENT OF A STEMMER FOR AFARAF TEXT.
Shi, T., Keneshloo, Y., Ramakrishnan, N., & Reddy, C. K. (2018). Neural Abstractive Text
Summarization with Sequence-to-Sequence Models: A Survey. ArXiv, February 2019.
Srivastava, A. (2020). Real Time Fake News Detection Using Machine Learning and NLP. June,
3679–3683.
Stahl, K. (2018). Fake news detection in online social media Problem Statement. May, 6.
https://leadingindia.ai/downloads/projects/SMA/sma_9.pdf
Stahl, K. (2019). Fake News Detector in Online Social Media. International Journal of
Engineering and Advanced Technology, 9(1S4), 58–60.
https://doi.org/10.35940/ijeat.a1089.1291s419
Talwar, S., Dhir, A., Kaur, P., & Zafar, N. (2019). Why do people share fake news ? Associations
between the dark side of social media use and fake news sharing behavior Journal of
Retailing and Consumer Services Why do people share fake news ? Associations between
the dark side of social media use and fak. September.
56 | P a g e
https://doi.org/10.1016/j.jretconser.2019.05.026
Tantiponganant, P., & Laksitamas, P. (2014). An analysis of the technology acceptance model in
understanding students’ behavioral intention to use university’s social media. In
Proceedings - 2014 IIAI 3rd International Conference on Advanced Applied Informatics,
IIAI-AAI 2014 (Issue September). https://doi.org/10.1109/IIAI-AAI.2014.14
Tesfaye, D. (2010b). Afaan Oromo Search Engine. A Thesis Submitted To the School of
Graduate Studies of the Addis Ababa University in Partial Fulfillment for the Degree of
Masters of Science in Computer Science, November.
Thota, A., Tilak, P., Ahluwalia, S., Lohia, N., Ahluwalia, S., & Lohia, N. (2018a). Fake News
Detection: A Deep Learning Approach. SMU Data Science Review, 1(3), 10.
https://scholar.smu.edu/datasciencereviewhttp://digitalrepository.smu.edu.Availableat:https:
//scholar.smu.edu/datasciencereview/vol1/iss3/10
Thota, A., Tilak, P., Ahluwalia, S., Lohia, N., Ahluwalia, S., & Lohia, N. (2018b). Fake News
Detection: A Deep Learning Approach. In SMU Data Science Review (Vol. 1, Issue 3).
https://scholar.smu.edu/datasciencereviewhttp://digitalrepository.smu.edu.Availableat:https:
//scholar.smu.edu/datasciencereview/vol1/iss3/10
Trung Tin, P. (2018). A Study on Deep Learning for Fake News Detection.
Tune, K. K., Varma, V., & Pingali, P. (2008). Evaluation of Oromo-English Cross-Language
Information Retrieval Evaluation of Oromo-English Cross-Language Information Retrieval.
June.
Vo, N., & Lee, K. (2019). Learning from fact-checkers: Analysis and generation of fact-checking
language. SIGIR 2019 - Proceedings of the 42nd International ACM SIGIR Conference on
Research and Development in Information Retrieval, 335–344.
57 | P a g e
https://doi.org/10.1145/3331184.3331248
Yulita, I. N., Fanany, M. I., & Arymuthy, A. M. (2017). Bi-directional Long Short-Term
Memory using Quantized data of Deep Belief Networks for Sleep Stage Classification.
Procedia Computer Science, 116, 530–538. https://doi.org/10.1016/j.procs.2017.10.042
Zaman, B., Justitia, A., Sani, K. N., & Purwanti, E. (2020). An Indonesian Hoax News Detection
System Using Reader Feedback and Naïve Bayes Algorithm. Cybernetics and Information
Technologies, 20(1), 82–94. https://doi.org/10.2478/cait-2020-0006
Zhang, J., Dong, B., & Yu, P. S. (2020). FakeDetector: Effective fake news detection with deep
diffusive neural network. Proceedings - International Conference on Data Engineering,
2020-April, 1826–1829. https://doi.org/10.1109/ICDE48307.2020.00180
Zhang, Z., & Luo, L. (2019). Hate speech detection: A solved problem? The challenging case of
long tail on Twitter. Semantic Web, 10(5), 925–945. https://doi.org/10.3233/SW-180338
Zhao, Z., Zhao, J., Sano, Y., Levy, O., Takayasu, H., Takayasu, M., Li, D., Wu, J., & Havlin, S.
(2020). Fake news propagates differently from real news even at early stages of spreading.
EPJ Data Science, 9(1). https://doi.org/10.1140/epjds/s13688-020-00224-z
58 | P a g e
Appendix
Appendix-1 Sample of Afan Oromo real news
ID Title Body Label
59 | P a g e
04 Humni tikaa Motummaa Pireezidaantiin Waldaa Maccaa 1
itiyophiya missense waldaa Tuulamaa Obbo Dirribii Damusee
maccaf tuulama duraani humnootii mootummaan reebichi
irratti reebicha irratti raawwatamuusaa BBC'tti
rawwataniiru bilbilaan himan.
60 | P a g e
04 Nannoo oromiyaa Manneen barnoota naannoole 0
oromiyaa keessas jiran hunda
keessatti afaan amariffa akka
fudhatamu ministerri barnoota
nannicha ifa godheera
61 | P a g e
tanaafi akka jechuun jechuu Illee Fi Moo
jechaan Osoo Odoo Akkum booda booddee Dura
Saniif tanaaf tanaafi tanaafuu itumallee waan hoggaa
Yeroo Akka ishee otuu akkasumas Malee Osoo
62 | P a g e