Abstract
We present CLIDSUM, a benchmark dataset towards building cross-lingual summarization systems on dialogue documents. It consists of 67k+ dialogue documents and 112k+ annotated summaries in different target languages. Based on the proposed CLIDSUM, we introduce two benchmark settings for supervised and semi-supervised scenarios, respectively. We then build various baseline systems in different paradigms (pipeline and end-to-end) and conduct extensive experiments on CLIDSUM to provide deeper analyses. Furthermore, we propose mDIALBART which extends mBART via further pre-training, where the multiple objectives help the pre-trained model capture the structural characteristics as well as key content in dialogues and the transformation from source to the target language. Experimental results show the superiority of mDIALBART, as an end-to-end model, outperforms strong pipeline models on CLIDSUM. Finally, we discuss specific challenges that current approaches faced with this task and give multiple promising directions for future research. We have released the dataset and code at https://github.com/krystalan/ClidSum.
Cite
CITATION STYLE
Wang, J., Meng, F., Lu, Z., Zheng, D., Li, Z., Qu, J., & Zhou, J. (2022). CLIDSUM: A Benchmark Dataset for Cross-Lingual Dialogue Summarization. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 7716–7729). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.526
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.