Abstract
We propose a novel data-augmentation technique for neural machine translation based on ROT-k ciphertexts. ROT-k is a simple letter substitution cipher that replaces a letter in the plaintext with the kth letter after it in the alphabet. We first generate multiple ROT-k ciphertexts using different values of k for the plaintext which is the source side of the parallel data. We then leverage this enciphered training data along with the original parallel data via multi-source training to improve neural machine translation. Our method, CipherDAug, uses a co-regularization-inspired training procedure, requires no external data sources other than the original training data, and uses a standard Transformer to outperform strong data augmentation techniques on several datasets by a significant margin. This technique combines easily with existing approaches to data augmentation, and yields particularly strong results in low-resource settings.
Cite
CITATION STYLE
Kambhatla, N., Born, L., & Sarkar, A. (2022). CipherDAug: Ciphertext based Data Augmentation for Neural Machine Translation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 201–218). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.17
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.