CIAug: Equipping Interpolative Augmentation with Curriculum Learning

1Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

Interpolative data augmentation has proven to be effective for NLP tasks. Despite its merits, the sample selection process in mixup is random, which might make it difficult for the model to generalize better and converge faster. We propose CIAug, a novel curriculum-based learning method that builds upon mixup. It leverages the relative position of samples in hyperbolic embedding space as a complexity measure to gradually mix up increasingly difficult and diverse samples along training. CIAug achieves state-of-the-art results over existing interpolative augmentation methods on 10 benchmark datasets across 4 languages in text classification and named-entity recognition tasks. It also converges and achieves benchmark F1 scores 3 times faster. We empirically analyze the various components of CIAug, and evaluate its robustness against adversarial attacks.

Cite

CITATION STYLE

APA

Sawhney, R., Soun, R., Pandit, S., Thakkar, M., Malaviya, S., & Pinter, Y. (2022). CIAug: Equipping Interpolative Augmentation with Curriculum Learning. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 1758–1764). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.127

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free