All Your Fake Detector Are Belong to Us: Evaluating Adversarial Robustness of Fake-News Detectors Under Black-Box Settings

37Citations
Citations of this article
57Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

With the hyperconnectivity and ubiquity of the Internet, the fake news problem now presents a greater threat than ever before. One promising solution for countering this threat is to leverage deep learning (DL)-based text classification methods for fake-news detection. However, since such methods have been shown to be vulnerable to adversarial attacks, the integrity and security of DL-based fake news classifiers are under question. Although many works study text classification under the adversarial threat, to the best of our knowledge, we do not find any work in literature that specifically analyzes the performance of DL-based fake-news detectors under adversarial settings. We bridge this gap by evaluating the performance of fake-news detectors under various configurations under black-box settings. In particular, we investigate the robustness of four different DL architectural choices-multilayer perceptron (MLP), convolutional neural network (CNN), recurrent neural network (RNN) and a recently proposed Hybrid CNN-RNN trained on three different state-of-the-art datasets-under different adversarial attacks (Text Bugger, Text Fooler, PWWS, and Deep Word Bug) implemented using the state-of-the-art NLP attack library, Text-Attack. Additionally, we explore how changing the detector complexity, the input sequence length, and the training loss affect the robustness of the learned model. Our experiments suggest that RNNs are robust as compared to other architectures. Further, we show that increasing the input sequence length generally increases the detector’s robustness. Our evaluations provide key insights to robustify fake-news detectors against adversarial attacks.

References Powered by Scopus

"Why should i trust you?" Explaining the predictions of any classifier

11970Citations
N/AReaders
Get full text

The limitations of deep learning in adversarial settings

3081Citations
N/AReaders
Get full text

Information credibility on Twitter

2093Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Functionality-Preserving Adversarial Machine Learning for Robust Classification in Cybersecurity and Intrusion Detection Domains: A Survey

39Citations
N/AReaders
Get full text

A systematic literature review and existing challenges toward fake news detection models

27Citations
N/AReaders
Get full text

Defense strategies for Adversarial Machine Learning: A survey

23Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Ali, H., Khan, M. S., Alghadhban, A., Alazmi, M., Alzamil, A., Al-Utaibi, K., & Qadir, J. (2021). All Your Fake Detector Are Belong to Us: Evaluating Adversarial Robustness of Fake-News Detectors Under Black-Box Settings. IEEE Access, 9, 81678–81692. https://doi.org/10.1109/ACCESS.2021.3085875

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 10

50%

Professor / Associate Prof. 4

20%

Researcher 4

20%

Lecturer / Post doc 2

10%

Readers' Discipline

Tooltip

Computer Science 13

65%

Social Sciences 3

15%

Engineering 3

15%

Arts and Humanities 1

5%

Save time finding and organizing research with Mendeley

Sign up for free