Professional Documents
Culture Documents
Table 3: Semantic similarity models. We present evaluation only for detection of the misinformation class, along
with an evaluation on relevance detection, i.e. on tweets that are either Misinformative or Informative.
Misinformative Informative
Model
P R F1 H@5 MRR P R F1 H@5 MRR
Trained on SNLI
Linear, Bag-of-Words 6.9 17.8 9.9 10.3 0.06 6.1 57.3 11.1 36.6 0.19
Linear, Avg. GloVe Embeddings 14.1 30.3 19.2 12.0 0.10 1.9 15.9 3.4 0.0 0.03
Sentence-BERT 8.0 9.0 8.5 9.5 0.10 2.6 26.8 4.7 0.0 0.02
Trained on MNLI
Linear, Bag-of-Words 8.8 52.7 15.1 9.0 0.08 10.1 48.8 16.7 16.5 0.15
Linear, Avg. GloVe Embeddings 16.7 64.3 26.5 3.9 0.05 2.7 23.8 4.8 17.7 0.09
Sentence-BERT 16.2 30.5 21.2 32.0 0.22 5.1 48.8 9.3 7.9 0.09
Table 4: NLI models. Classification and ranking evaluation metrics for Misinformative or Informative classes.
Giovanni Luca Ciampaglia, Prashant Shiralkar, Luis M Kai Shu, Suhang Wang, and Huan Liu. 2019. Beyond
Rocha, Johan Bollen, Filippo Menczer, and news contents: The role of social context for fake
Alessandro Flammini. 2015. Computational fact news detection. In Proceedings of the Twelfth ACM
checking from knowledge networks. PloS one, International Conference on Web Search and Data
10(6):e0128193. Mining, pages 312–320.
Suchin Gururangan, Ana Marasovi, Swabha TechCrunch. 2020. Facebook, reddit, google, linkedin,
Swayamdipta, Kyle Lo, Iz Beltagy, Doug Downey, microsoft, twitter and youtube issue joint statement
and Noah A. Smith. 2020. Don’t stop pretraining: on misinformation. [Accessed on June 30, 2020].
Adapt language models to domains and tasks. In
Proceedings of ACL. James Thorne, Mingjie Chen, Giorgos Myrianthous,
Jiashu Pu, Xiaoxuan Wang, and Andreas Vlachos.
Hamid Karimi, Proteek Roy, Sari Saba-Sadiya, and 2017. Fake news stance detection using stacked en-
Jiliang Tang. 2018. Multi-source multi-class fake semble of classifiers. In Proceedings of the 2017
news detection. In Proceedings of the 27th Inter- EMNLP Workshop: Natural Language Processing
national Conference on Computational Linguistics, meets Journalism, pages 80–83, Copenhagen, Den-
pages 1546–1557. mark. Association for Computational Linguistics.
Srijan Kumar, Robert West, and Jure Leskovec. 2016. James Thorne and Andreas Vlachos. 2018. Automated
Disinformation on the web: Impact, characteristics, fact checking: Task formulations, methods and fu-
and detection of wikipedia hoaxes. In Proceedings ture directions. In Proceedings of the 27th Inter-
of the 25th international conference on World Wide national Conference on Computational Linguistics,
Web, pages 591–602. pages 3346–3359, Santa Fe, New Mexico, USA. As-
sociation for Computational Linguistics.
Yaliang Li, Qi Li, Jing Gao, Lu Su, Bo Zhao, Wei Fan,
and Jiawei Han. 2015. On the discovery of evolv- James Thorne, Andreas Vlachos, Oana Cocarascu,
ing truth. In Proceedings of the 21th ACM SIGKDD Christos Christodoulopoulos, and Arpit Mittal. 2018.
International Conference on Knowledge Discovery The fact extraction and verification (fever) shared
and Data Mining, pages 675–684. task. arXiv preprint arXiv:1811.10971.
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Man- Sebastian Tschiatschek, Adish Singla, Manuel
dar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Gomez Rodriguez, Arpit Merchant, and Andreas
Luke Zettlemoyer, and Veselin Stoyanov. 2019. Krause. 2018. Fake news detection in social
networks via crowd signals. In Companion Pro- A Reliability Codes
ceedings of the The Web Conference 2018, pages
517–524.
• A score of 5 were sources that clearly debunked
Svitlana Volkova, Kyle Shaffer, Jin Yea Jang, and
Nathan Hodas. 2017. Separating facts from fiction: the paired misinformation, cited evidence, and
Linguistic models to classify suspicious and trusted were from a reputable source or well-known fact-
news posts on twitter. In Proceedings of the 55th An- check website (e.g. CDC, Snopes).
nual Meeting of the Association for Computational • A score of 4 were sources that debunked the
Linguistics (Volume 2: Short Papers), pages 647–
653.
paired misinformation and were from a reliable
source (e.g. News sites).
William Yang Wang. 2017. ” liar, liar pants on fire”: • A score of 3 were sources that refuted the misin-
A new benchmark dataset for fake news detection. formation but were not a well-known source or
arXiv preprint arXiv:1705.00648.
contained a lot of filler in their article not about
Wei Wei and Xiaojun Wan. 2017. Learning to identify the misinformation.
ambiguous and misleading news headlines. arXiv • A score of 2 were sources that labeled the mis-
preprint arXiv:1705.06031.
information as false or untrue but did not really
WHO. 2020. Novel coronavirus(2019-ncov) situation provide evidence.
report - 13. [Accessed on June 30, 2020]. • A score of 1 were sources that did not refute the
Wikipedia. 2020. Misinformation related to the misinformation or were more descriptive.
covid-19 pandemic: https://en.wikipedia.org/wiki/ • A score of 0 were sources that did not refute the
Misinformation related to the COVID-19 pandemic. misinformation and may actually support it.
• A score of NA for sources not in English.
Adina Williams, Nikita Nangia, and Samuel Bowman.
2018. A broad-coverage challenge corpus for sen-
tence understanding through inference. In Proceed-
ings of the 2018 Conference of the North American
Chapter of the Association for Computational Lin-
guistics: Human Language Technologies, Volume
1 (Long Papers), pages 1112–1122. Association for
Computational Linguistics.
Kai-Chou Yang, Timothy Niven, and Hung-Yu Kao.
2019. Fake news detection as natural language in-
ference. arXiv preprint arXiv:1907.07347.
Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q
Weinberger, and Yoav Artzi. 2019. Bertscore: Eval-
uating text generation with bert. arXiv preprint
arXiv:1904.09675.