A dataset and evaluation metrics for abstractive compression of sentences and short paragraphs

35Citations
Citations of this article
127Readers
Mendeley users who have this article in their library.

Abstract

We introduce a manually-created, multi-reference dataset for abstractive sentence and short paragraph compression. First, we examine the impact of single- and multi-sentence level editing operations on human compression quality as found in this corpus. We observe that substitution and rephrasing operations are more meaning preserving than other operations, and that compressing in context improves quality. Second, we systematically explore the correlations between automatic evaluation metrics and human judgments of meaning preservation and grammaticality in the compression task, and analyze the impact of the linguistic units used and precision versus recall measures on the quality of the metrics. Multi-reference evaluation metrics are shown to offer significant advantage over single reference-based metrics.

Cite

CITATION STYLE

APA

Toutanova, K., Tran, K. M., Brockett, C., & Amershi, S. (2016). A dataset and evaluation metrics for abstractive compression of sentences and short paragraphs. In EMNLP 2016 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 340–350). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d16-1033

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free