Neural abstractive summarization with structural attention

12Citations
Citations of this article
49Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Attentional, RNN-based encoder-decoder architectures have achieved impressive performance on abstractive summarization of news articles. However, these methods fail to account for long term dependencies within the sentences of a document. This problem is exacerbated in multi-document summarization tasks such as summarizing the popular opinion in threads present in community question answering (CQA) websites such as Yahoo! Answers and Quora. These threads contain answers which often overlap or contradict each other. In this work, we present a hierarchical encoder based on structural attention to model such inter-sentence and inter-document dependencies. We set the popular pointer-generator architecture and some of the architectures derived from it as our baselines and show that they fail to generate good summaries in a multi-document setting. We further illustrate that our proposed model achieves significant improvement over the baselines in both single and multi-document summarization settings - in the former setting, it beats the best baseline by 1.31 and 7.8 ROUGE-1 points on CNN and CQA datasets, respectively; in the latter setting, the performance is further improved by 1.6 ROUGE-1 points on the CQA dataset.

Cite

CITATION STYLE

APA

Chowdhury, T., Kumar, S., & Chakraborty, T. (2020). Neural abstractive summarization with structural attention. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 3716–3722). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2020/514

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free