A Hierarchical Encoding-Decoding Scheme for Abstractive Multi-document Summarization

5Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Pre-trained language models (PLMs) have achieved outstanding achievements in abstractive single-document summarization (SDS). However, such benefits may not fully extend to multi-document summarization (MDS), where the handling of cross-document information is more complex. Previous works either design new MDS architectures or apply PLMs bluntly with concatenated source documents as a reformulated SDS task. While the former does not utilize previous pre-training efforts and may not generalize well across different domains, the latter may not sufficiently attend to the intricate cross-document relationships unique to MDS tasks. Instead, we enforce hierarchy on both the encoder and decoder to better utilize a PLM to facilitate multi-document interactions for the MDS task. Across 10 MDS benchmarks from various domains, our method outperforms or is competitive with the previous best models, including those with additional MDS pre-training or with more parameters. It outperforms its corresponding PLM backbone by up to 3 ROUGE-L and is favored by humans.

Cite

CITATION STYLE

APA

Shen, C., Cheng, L., Nguyen, X. P., You, Y., & Bing, L. (2023). A Hierarchical Encoding-Decoding Scheme for Abstractive Multi-document Summarization. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 5872–5887). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.391

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free