Fact-level Extractive Summarization with Hierarchical Graph Mask on BERT

10Citations
Citations of this article
94Readers
Mendeley users who have this article in their library.

Abstract

Most current extractive summarization models generate summaries by selecting salient sentences. However, one of the problems with sentence-level extractive summarization is that there exists a gap between the human-written gold summary and the oracle sentence labels. In this paper, we propose to extract fact-level semantic units for better extractive summarization. We also introduce a hierarchical structure, which incorporates the multi-level of granularities of the textual information into the model. In addition, we incorporate our model with BERT using a hierarchical graph mask. This allows us to combine BERT’s ability in natural language understanding and the structural information without increasing the scale of the model. Experiments on the CNN/DaliyMail dataset show that our model achieves state-of-the-art results.

Cite

CITATION STYLE

APA

Yuan, R., Wang, Z., & Li, W. (2020). Fact-level Extractive Summarization with Hierarchical Graph Mask on BERT. In COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference (pp. 5629–5639). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.coling-main.493

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free