Contrastive Representation Learning for Exemplar-Guided Paraphrase Generation

10Citations
Citations of this article
56Readers
Mendeley users who have this article in their library.

Abstract

Exemplar-Guided Paraphrase Generation (EGPG) aims to generate a target sentence which conforms to the style of the given exemplar while encapsulating the content information of the source sentence. In this paper, we propose a new method with the goal of learning a better representation of the style and the content. This method is mainly motivated by the recent success of contrastive learning which has demonstrated its power in unsupervised feature extraction tasks. The idea is to design two contrastive losses with respect to the content and the style by considering two problem characteristics during training. One characteristic is that the target sentence shares the same content with the source sentence, and the second characteristic is that the target sentence shares the same style with the exemplar. These two contrastive losses are incorporated into the general encoder-decoder paradigm. Experiments on two datasets, namely QQP-Pos and ParaNMT, demonstrate the effectiveness of our proposed constrastive losses. The code is available at https://github.com/LHRYANG/CRL_EGPG.

Cite

CITATION STYLE

APA

Yang, H., Lam, W., & Li, P. (2021). Contrastive Representation Learning for Exemplar-Guided Paraphrase Generation. In Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 (pp. 4754–4761). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-emnlp.409

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free