Context-aware translation can be achieved by processing a concatenation of consecutive sentences with the standard Transformer architecture. This paper investigates the intuitive idea of providing the model with explicit information about the position of the sentences contained in the concatenation window. We compare various methods to encode sentence positions into token representations, including novel methods. Our results show that the Transformer benefits from certain sentence position encodings methods on En→Ru, if trained with a context-discounted loss (Lupo et al., 2022b). However, the same benefits are not observed on En→De. Further empirical efforts are necessary to define the conditions under which the proposed approach is beneficial.
Mendeley helps you to discover research relevant for your work.
CITATION STYLE
Lupo, L., Dinarelli, M., & Besacier, L. (2023). Encoding Sentence Position in Context-Aware Neural Machine Translation with Concatenation. In ACL 2023 - 4th Workshop on Insights from Negative Results in NLP, Proceedings (pp. 33–44). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.insights-1.4