Global Encoding for Long Chinese Text Summarization

18Citations
Citations of this article
26Readers
Mendeley users who have this article in their library.

Abstract

Text summarization is one of the significant tasks of natural language processing, which automatically converts text into a summary. Some summarization systems, for short/long English, and short Chinese text, benefit from advances in the neural encoder-decoder model because of the availability of large datasets. However, the long Chinese text summarization research has been limited to datasets of a couple of hundred instances. This article aims to explore the long Chinese text summarization task. To begin with, we construct a first large-scale, long Chinese text summarization corpus, the Long Chinese Summarization of Police Inquiry Record Text (LCSPIRT). Based on this corpus, we propose a sequence-to-sequence (Seq2Seq) model that incorporates a global encoding process with an attention mechanism. Our model achieves a competitive result on the LCSPIRT corpus compared with several benchmark methods.

Cite

CITATION STYLE

APA

Xi, X., Pi, Z., & Zhou, G. (2020). Global Encoding for Long Chinese Text Summarization. ACM Transactions on Asian and Low-Resource Language Information Processing, 19(6). https://doi.org/10.1145/3407911

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free