A Hierarchical Representation Model Based on Longformer and Transformer for Extractive Summarization

6Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

Automatic text summarization is a method used to compress documents while preserving the main idea of the original text, including extractive summarization and abstractive summarization. Extractive text summarization extracts important sentences from the original document to serve as the summary. The document representation method is crucial for the quality of the generated summarization. To effectively represent the document, we propose a hierarchical document representation model Long-Trans-Extr for Extractive Summarization, which uses Longformer as the sentence encoder and Transformer as the document encoder. The advantage of Longformer as sentence encoder is that the model can input long document up to 4096 tokens with adding relative a little calculation. The proposed model Long-Trans-Extr is evaluated on three benchmark datasets: CNN (Cable News Network), DailyMail, and the combined CNN/DailyMail. It achieves 43.78 (Rouge-1) and 39.71 (Rouge-L) on CNN/DailyMail and 33.75 (Rouge-1), 13.11 (Rouge-2), and 30.44 (Rouge-L) on the CNN datasets. They are very competitive results, and furthermore, they show that our model has better performance on long documents, such as the CNN corpus.

References Powered by Scopus

Long Short-Term Memory

78506Citations
N/AReaders
Get full text

Learning phrase representations using RNN encoder-decoder for statistical machine translation

11829Citations
N/AReaders
Get full text

Get to the point: Summarization with pointer-generator networks

2646Citations
N/AReaders
Get full text

Cited by Powered by Scopus

A Comprehensive Analysis of Indian Legal Documents Summarization Techniques

10Citations
N/AReaders
Get full text

A systematic literature review of deep learning-based text summarization: Techniques, input representation, training strategies, mechanisms, datasets, evaluation, and challenges

6Citations
N/AReaders
Get full text

Implementation of Preprocessing in Text Summarization Techniques for Indonesian Language Documents Using the Flax T5 Approach

1Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Yang, S., Zhang, S., Fang, M., Yang, F., & Liu, S. (2022). A Hierarchical Representation Model Based on Longformer and Transformer for Extractive Summarization. Electronics (Switzerland), 11(11). https://doi.org/10.3390/electronics11111706

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 4

80%

Researcher 1

20%

Readers' Discipline

Tooltip

Computer Science 3

75%

Physics and Astronomy 1

25%

Save time finding and organizing research with Mendeley

Sign up for free