We present a novel technique for zero-shot paraphrase generation. The key contribution is an end-to-end multilingual paraphrasing model that is trained using translated parallel corpora to generate paraphrases into “meaning spaces” – replacing the final softmax layer with word embeddings. This architectural modification, plus a training procedure that incorporates an autoencoding objective, enables effective parameter sharing across languages for more fluent monolingual rewriting, and facilitates fluency and diversity in generation. Our continuous-output paraphrase generation models outperform zero-shot paraphrasing baselines, when evaluated on two languages using a battery of computational metrics as well as in human assessment.
CITATION STYLE
Jegadeesan, M., Kumar, S., Wieting, J., & Tsvetkov, Y. (2021). Improving the Diversity of Unsupervised Paraphrasing with Embedding Outputs. In MRL 2021 - 1st Workshop on Multilingual Representation Learning, Proceedings of the Conference (pp. 166–175). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.mrl-1.15
Mendeley helps you to discover research relevant for your work.