Increasing robustness to spurious correlations using forgettable examples

38Citations
Citations of this article
80Readers
Mendeley users who have this article in their library.

Abstract

Neural NLP models tend to rely on spurious correlations between labels and input features to perform their tasks. Minority examples, i.e., examples that contradict the spurious correlations present in the majority of data points, have been shown to increase the out-of-distribution generalization of pre-trained language models. In this paper, we first propose using example forgetting to find minority examples without prior knowledge of the spurious correlations present in the dataset. Forgettable examples are instances either learned and then forgotten during training or never learned. We empirically show how these examples are related to minorities in our training sets. Then, we introduce a new approach to robustify models by fine-tuning our models twice, first on the full training data and second on the minorities only. We obtain substantial improvements in out-of-distribution generalization when applying our approach to the MNLI, QQP, and FEVER datasets.

Cite

CITATION STYLE

APA

Yaghoobzadeh, Y., Mehri, S., des Combes, R. T., Hazen, T. J., & Sordoni, A. (2021). Increasing robustness to spurious correlations using forgettable examples. In EACL 2021 - 16th Conference of the European Chapter of the Association for Computational Linguistics, Proceedings of the Conference (pp. 3319–3332). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.eacl-main.291

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free