CLIDSUM: A Benchmark Dataset for Cross-Lingual Dialogue Summarization

23Citations
Citations of this article
36Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present CLIDSUM, a benchmark dataset towards building cross-lingual summarization systems on dialogue documents. It consists of 67k+ dialogue documents and 112k+ annotated summaries in different target languages. Based on the proposed CLIDSUM, we introduce two benchmark settings for supervised and semi-supervised scenarios, respectively. We then build various baseline systems in different paradigms (pipeline and end-to-end) and conduct extensive experiments on CLIDSUM to provide deeper analyses. Furthermore, we propose mDIALBART which extends mBART via further pre-training, where the multiple objectives help the pre-trained model capture the structural characteristics as well as key content in dialogues and the transformation from source to the target language. Experimental results show the superiority of mDIALBART, as an end-to-end model, outperforms strong pipeline models on CLIDSUM. Finally, we discuss specific challenges that current approaches faced with this task and give multiple promising directions for future research. We have released the dataset and code at https://github.com/krystalan/ClidSum.

Cite

CITATION STYLE

APA

Wang, J., Meng, F., Lu, Z., Zheng, D., Li, Z., Qu, J., & Zhou, J. (2022). CLIDSUM: A Benchmark Dataset for Cross-Lingual Dialogue Summarization. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 7716–7729). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.526

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free