R-AT: Regularized Adversarial Training for Natural Language Understanding

6Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

Abstract

Currently, adversarial training has become a popular and powerful regularization method in the natural language domain. In this paper, we Regularized Adversarial Training (R-AT) via dropout, which forces the output probability distributions of different sub-models generated by dropout to be consistent under the same adversarial samples. Specifically, we generate adversarial samples by perturbing the word embeddings. For each adversarial sample fed to the model, R-AT minimizes both the adversarial risk and the bidirectional KL-divergence between the adversarial output distributions of two sub-models sampled by dropout. Through extensive experiments on 13 public natural language understanding datasets, we found that RAT has improvements for many models (e.g., rnn-based, cnn-based, and transformer-based models). For the GLUE benchmark, when RAT is only applied to the fine-tuning stage, it is able to improve the overall test score of the BERT-base model from 78.3 to 79.6 and the RoBERTa-large model from 88.1 to 88.6. Theoretical analysis reveals that R-AT has potential gradient regularization during the training process. Furthermore, R-AT can reduce the inconsistency between training and testing of models with dropout.

Cite

CITATION STYLE

APA

Ni, S., Li, J., & Kao, H. Y. (2022). R-AT: Regularized Adversarial Training for Natural Language Understanding. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 6456–6469). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.480

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free