Adv-OLM: Generating textual adversaries via OLM

3Citations
Citations of this article
70Readers
Mendeley users who have this article in their library.

Abstract

Deep learning models are susceptible to adversarial examples that have imperceptible perturbations in the original input, resulting in adversarial attacks against these models. Analysis of these attacks on the state of the art transformers in NLP can help improve the robustness of these models against such adversarial inputs. In this paper, we present Adv-OLM, a black-box attack method that adapts the idea of Occlusion and Language Models (OLM) to the current state of the art attack methods. OLM is used to rank words of a sentence, which are later substituted using word replacement strategies. We experimentally show that our approach outperforms other attack methods for several text classification tasks.

Cite

CITATION STYLE

APA

Malik, V., Bhat, A., & Modi, A. (2021). Adv-OLM: Generating textual adversaries via OLM. In EACL 2021 - 16th Conference of the European Chapter of the Association for Computational Linguistics, Proceedings of the Conference (pp. 841–849). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.eacl-main.71

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free