Abstract
Most current extractive summarization models generate summaries by selecting salient sentences. However, one of the problems with sentence-level extractive summarization is that there exists a gap between the human-written gold summary and the oracle sentence labels. In this paper, we propose to extract fact-level semantic units for better extractive summarization. We also introduce a hierarchical structure, which incorporates the multi-level of granularities of the textual information into the model. In addition, we incorporate our model with BERT using a hierarchical graph mask. This allows us to combine BERT’s ability in natural language understanding and the structural information without increasing the scale of the model. Experiments on the CNN/DaliyMail dataset show that our model achieves state-of-the-art results.
Cite
CITATION STYLE
Yuan, R., Wang, Z., & Li, W. (2020). Fact-level Extractive Summarization with Hierarchical Graph Mask on BERT. In COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference (pp. 5629–5639). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.coling-main.493
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.