Summary++: Summarizing Chinese News Articles with Attention

1Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present Summary++, the model that competed in NLPCC2018’s Summary task. In this paper, we describe in detail of the task, our model, the results and other aspects during our experiments. The task is News article summarization in Chinese, where one sentence is generated per article. We use a neural encoder decoder attention model with pointer generator network, and modify it to focus on words attented to rather than words predicted. Our model archive second place in the task with a score of 0.285. The highlights of our model is that it run at character level, no extra features (e.g. part of speech, dependency structure) were used and very little preprocessing were done.

Cite

CITATION STYLE

APA

Zhao, J., Chung, T. L., Xu, B., & Jiang, M. (2018). Summary++: Summarizing Chinese News Articles with Attention. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11109 LNAI, pp. 27–37). Springer Verlag. https://doi.org/10.1007/978-3-319-99501-4_3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free