This paper describes our submitted text-to-text Simultaneous translation (ST) system, which won the second place in the Chinese→English streaming translation task of AutoSimTrans 2022. Our baseline system is a BPE-based Transformer model trained with the PaddlePaddle framework. In our experiments, we employ data synthesis and ensemble approaches to enhance the base model. In order to bridge the gap between general domain and spoken domain, we select in-domain data from a general corpus and mix them with a spoken corpus for mixed fine-tuning. Finally, we adopt a fixed wait-k policy to transfer our full-sentence translation model to simultaneous translation model. Experiments on the development data show that our system outperforms the baseline system.
CITATION STYLE
Zhu, J., & Yu, J. (2022). USST’s System for AutoSimTrans 2022. In AutoSimTrans 2022 - Automatic Simultaneous Translation Challenges, Recent Advances, and Future Directions, Proceedings of the 3rd Workshop (pp. 43–49). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.autosimtrans-1.7
Mendeley helps you to discover research relevant for your work.