Abstract
A recent study by Feldman (2020) proposed a long-tail theory to explain the memorization behavior of deep learning models. However, memorization has not been empirically verified in the context of NLP, a gap addressed by this work. In this paper, we use three different NLP tasks to check if the long-tail theory holds. Our experiments demonstrate that top-ranked memorized training instances are likely atypical, and removing the top-memorized training instances leads to a more serious drop in test accuracy compared with removing training instances randomly. Furthermore, we develop an attribution method to better understand why a training instance is memorized. We empirically show that our memorization attribution method is faithful and share our interesting finding that the top-memorized parts of a training instance tend to be features negatively correlated with the class label.
Cite
CITATION STYLE
Zheng, X., & Jiang, J. (2022). An Empirical Study of Memorization in NLP. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 6265–6278). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.434
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.