Improving the Diversity of Unsupervised Paraphrasing with Embedding Outputs

0Citations
Citations of this article
45Readers
Mendeley users who have this article in their library.

Abstract

We present a novel technique for zero-shot paraphrase generation. The key contribution is an end-to-end multilingual paraphrasing model that is trained using translated parallel corpora to generate paraphrases into “meaning spaces” – replacing the final softmax layer with word embeddings. This architectural modification, plus a training procedure that incorporates an autoencoding objective, enables effective parameter sharing across languages for more fluent monolingual rewriting, and facilitates fluency and diversity in generation. Our continuous-output paraphrase generation models outperform zero-shot paraphrasing baselines, when evaluated on two languages using a battery of computational metrics as well as in human assessment.

Cite

CITATION STYLE

APA

Jegadeesan, M., Kumar, S., Wieting, J., & Tsvetkov, Y. (2021). Improving the Diversity of Unsupervised Paraphrasing with Embedding Outputs. In MRL 2021 - 1st Workshop on Multilingual Representation Learning, Proceedings of the Conference (pp. 166–175). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.mrl-1.15

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free