Token-Level Self-Evolution Training for Sequence-to-Sequence Learning

6Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Adaptive training approaches, widely used in sequence-to-sequence models, commonly reweigh the losses of different target tokens based on priors, e.g. word frequency. However, most of them do not consider the variation of learning difficulty in different training steps, and overly emphasize the learning of difficult one-hot labels, making the learning deterministic and sub-optimal. In response, we present Token-Level Self-Evolution Training (SE), a simple and effective dynamic training method to fully and wisely exploit the knowledge from data. SE focuses on dynamically learning the under-explored tokens for each forward pass and adaptively regularizes the training by introducing a novel token-specific label smoothing approach. Empirically, SE yields consistent and significant improvements in three tasks, i.e. machine translation, summarization, and grammatical error correction. Encouragingly, we achieve averaging +0.93 BLEU improvement on three machine translation tasks. Analyses confirm that, besides improving lexical accuracy, SE enhances generation diversity and model generalization.

Cite

CITATION STYLE

APA

Peng, K., Ding, L., Zhong, Q., Ouyang, Y., Rong, W., Xiong, Z., & Tao, D. (2023). Token-Level Self-Evolution Training for Sequence-to-Sequence Learning. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 2, pp. 841–850). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-short.73

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free