Learning to Consider Relevance and Redundancy Dynamically for Abstractive Multi-document Summarization

2Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

As one of the most essential tasks for information aggregation, multi-document summarization is faced with information redundancy of source document clusters. Recent works have attempted to avoid redundancy while generating summaries. Most state-of-the-art multi-document summarization systems are either extractive or abstractive with an external extractive model. In this paper, we propose an end-to-end abstractive model based on Transformer to generate summaries, considering relevance and redundancy dynamically and jointly. Specifically, we employ sentence masks and design a sentence-level transformer layer for learning sentence representations in a hierarchical manner. Then we use a dynamic Max Marginal Relevance (MMR) model to discern summary-worthy sentences and modify the encoder-decoder attention. We also utilize the pointer mechanism, taking the mean attention of all transformer heads as the probability to copy words from the source text. Experimental results demonstrate that our proposed model outperforms several strong baselines. We also conduct ablation studies to verify the effectiveness of our key mechanisms.

Cite

CITATION STYLE

APA

Liu, Y., Fan, X., Zhou, J., He, C., & Liu, G. (2020). Learning to Consider Relevance and Redundancy Dynamically for Abstractive Multi-document Summarization. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12430 LNAI, pp. 482–493). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-60450-9_38

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free