Long warm-up and self-training: Training strategies of NICT-2 NMT system at WAT-2019

2Citations
Citations of this article
63Readers
Mendeley users who have this article in their library.

Abstract

This paper describes the NICT-2 neural machine translation system at the 6th Workshop on Asian Translation. This system employs the standard Transformer model but features the following two characteristics. One is the long warm-up strategy, which performs a longer warm-up of the learning rate at the start of the training than conventional approaches. Another is that the system introduces self-training approaches based on multiple back-translations generated by sampling. We participated in three tasks-ASPEC.enja, ASPEC.ja-en, and TDDC.ja-en-using this system.

Cite

CITATION STYLE

APA

Imamura, K., & Sumita, E. (2021). Long warm-up and self-training: Training strategies of NICT-2 NMT system at WAT-2019. In WAT@EMNLP-IJCNLP 2019 - 6th Workshop on Asian Translation, Proceedings (pp. 141–146). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d19-5217

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free