An Empirical Study of Memorization in NLP

15Citations
Citations of this article
67Readers
Mendeley users who have this article in their library.

Abstract

A recent study by Feldman (2020) proposed a long-tail theory to explain the memorization behavior of deep learning models. However, memorization has not been empirically verified in the context of NLP, a gap addressed by this work. In this paper, we use three different NLP tasks to check if the long-tail theory holds. Our experiments demonstrate that top-ranked memorized training instances are likely atypical, and removing the top-memorized training instances leads to a more serious drop in test accuracy compared with removing training instances randomly. Furthermore, we develop an attribution method to better understand why a training instance is memorized. We empirically show that our memorization attribution method is faithful and share our interesting finding that the top-memorized parts of a training instance tend to be features negatively correlated with the class label.

Cite

CITATION STYLE

APA

Zheng, X., & Jiang, J. (2022). An Empirical Study of Memorization in NLP. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 6265–6278). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.434

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free