Using Random Perturbations to Mitigate Adversarial Attacks on NLP Models

10Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

Deep learning models have excelled in solving many problems in Natural Language Processing, but are susceptible to extensive vulnerabilities. We offer a solution to this vulnerability by using random perturbations such as spelling correction, synonym substitution, or dropping the word. These perturbations are applied to random words in random sentences to defend NLP models against adversarial attacks. Our defense methods are successful in returning attacked models to their original accuracy within statistical significance.

Cite

CITATION STYLE

APA

Swenor, A. (2022). Using Random Perturbations to Mitigate Adversarial Attacks on NLP Models. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 13142–13143). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i11.21707

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free