This paper describes the NICT-2 neural machine translation system at the 6th Workshop on Asian Translation. This system employs the standard Transformer model but features the following two characteristics. One is the long warm-up strategy, which performs a longer warm-up of the learning rate at the start of the training than conventional approaches. Another is that the system introduces self-training approaches based on multiple back-translations generated by sampling. We participated in three tasks-ASPEC.enja, ASPEC.ja-en, and TDDC.ja-en-using this system.
CITATION STYLE
Imamura, K., & Sumita, E. (2021). Long warm-up and self-training: Training strategies of NICT-2 NMT system at WAT-2019. In WAT@EMNLP-IJCNLP 2019 - 6th Workshop on Asian Translation, Proceedings (pp. 141–146). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d19-5217
Mendeley helps you to discover research relevant for your work.