We evaluated a range of neural machine translation techniques developed specifically for low-resource scenarios. Unsuccessfully. In the end, we submitted two runs: (i) a standard phrase-based model, and (ii) a random babbling baseline using character trigrams. We found that it was surprisingly hard to beat (i), in spite of this model being, in theory, a bad fit for polysynthetic languages; and more interestingly, that (ii) was better than several of the submitted systems, highlighting how difficult low-resource machine translation for polysynthetic languages is.
CITATION STYLE
Bollmann, M., Aralikatte, R., Bello, H. R. M., Hershcovich, D., de Lhoneux, M., & Søgaard, A. (2021). Moses and the Character-Based Random Babbling Baseline: CoAStaL at AmericasNLP 2021 Shared Task. In Proceedings of the 1st Workshop on Natural Language Processing for Indigenous Languages of the Americas, AmericasNLP 2021 (pp. 248–254). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.americasnlp-1.28
Mendeley helps you to discover research relevant for your work.