Global Encoding for Long Chinese Text Summarization

20Citations
Citations of this article
26Readers
Mendeley users who have this article in their library.

Abstract

Text summarization is one of the significant tasks of natural language processing, which automatically converts text into a summary. Some summarization systems, for short/long English, and short Chinese text, benefit from advances in the neural encoder-decoder model because of the availability of large datasets. However, the long Chinese text summarization research has been limited to datasets of a couple of hundred instances. This article aims to explore the long Chinese text summarization task. To begin with, we construct a first large-scale, long Chinese text summarization corpus, the Long Chinese Summarization of Police Inquiry Record Text (LCSPIRT). Based on this corpus, we propose a sequence-to-sequence (Seq2Seq) model that incorporates a global encoding process with an attention mechanism. Our model achieves a competitive result on the LCSPIRT corpus compared with several benchmark methods.

References Powered by Scopus

Rethinking the Inception Architecture for Computer Vision

24509Citations
N/AReaders
Get full text

Get to the point: Summarization with pointer-generator networks

2646Citations
N/AReaders
Get full text

LexRank: Graph-based lexical centrality as salience in text summarization

2445Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Few-Shot Fine-Tuning SOTA Summarization Models for Medical Dialogues

11Citations
N/AReaders
Get full text

Extractive summarization of Malayalam documents using latent Dirichlet allocation: An experience

6Citations
N/AReaders
Get full text

Learning to Summarize Chinese Radiology Findings with a Pre-Trained Encoder

6Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Xi, X., Pi, Z., & Zhou, G. (2020). Global Encoding for Long Chinese Text Summarization. ACM Transactions on Asian and Low-Resource Language Information Processing, 19(6). https://doi.org/10.1145/3407911

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 5

50%

Lecturer / Post doc 2

20%

Researcher 2

20%

Professor / Associate Prof. 1

10%

Readers' Discipline

Tooltip

Computer Science 11

85%

Medicine and Dentistry 1

8%

Chemistry 1

8%

Save time finding and organizing research with Mendeley

Sign up for free