Rewards with Negative Examples for Reinforced Topic-Focused Abstractive Summarization

2Citations
Citations of this article
36Readers
Mendeley users who have this article in their library.

Abstract

We consider the problem of topic-focused abstractive summarization, where the goal is to generate an abstractive summary focused on a particular topic, a phrase of one or multiple words. We hypothesize that the task of generating topic-focused summaries can be improved by showing the model what it must not focus on. We introduce a deep reinforcement learning approach to topic-focused abstractive summarization, trained on rewards with a novel negative example baseline. We define the input in this problem as the source text preceded by the topic. We adapt the CNN-Daily Mail and New York Times summarization datasets for this task. We then show through experiments on existing rewards that the use of a negative example baseline can outperform the use of a self-critical baseline, in ROUGE, BERTSCORE, and human evaluation metrics.

Cite

CITATION STYLE

APA

Mrini, K., Liu, C., & Dreyer, M. (2021). Rewards with Negative Examples for Reinforced Topic-Focused Abstractive Summarization. In 3rd Workshop on New Frontiers in Summarization, NewSum 2021 - Workshop Proceedings (pp. 33–38). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.newsum-1.4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free