Generative Pretraining for Paraphrase Evaluation

9Citations
Citations of this article
41Readers
Mendeley users who have this article in their library.

Abstract

We introduce ParaBLEU, a paraphrase representation learning model and evaluation metric for text generation. Unlike previous approaches, ParaBLEU learns to understand paraphrasis using generative conditioning as a pretraining objective. ParaBLEU correlates more strongly with human judgements than existing metrics, obtaining new state-of-the-art results on the 2017 WMT Metrics Shared Task. We show that our model is robust to data scarcity, exceeding previous state-of-the-art performance using only 50% of the available training data and surpassing BLEU, ROUGE and METEOR with only 40 labelled examples. Finally, we demonstrate that ParaBLEU can be used to conditionally generate novel paraphrases from a single demonstration, which we use to confirm our hypothesis that it learns abstract, generalized paraphrase representations.

Cite

CITATION STYLE

APA

Weston, J., Lenain, R., Meepegama, U., & Fristed, E. (2022). Generative Pretraining for Paraphrase Evaluation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 4052–4073). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.280

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free