Text summarization is one of the significant tasks of natural language processing, which automatically converts text into a summary. Some summarization systems, for short/long English, and short Chinese text, benefit from advances in the neural encoder-decoder model because of the availability of large datasets. However, the long Chinese text summarization research has been limited to datasets of a couple of hundred instances. This article aims to explore the long Chinese text summarization task. To begin with, we construct a first large-scale, long Chinese text summarization corpus, the Long Chinese Summarization of Police Inquiry Record Text (LCSPIRT). Based on this corpus, we propose a sequence-to-sequence (Seq2Seq) model that incorporates a global encoding process with an attention mechanism. Our model achieves a competitive result on the LCSPIRT corpus compared with several benchmark methods.
CITATION STYLE
Xi, X., Pi, Z., & Zhou, G. (2020). Global Encoding for Long Chinese Text Summarization. ACM Transactions on Asian and Low-Resource Language Information Processing, 19(6). https://doi.org/10.1145/3407911
Mendeley helps you to discover research relevant for your work.