Paraphrases refer to texts that convey the same meaning with different expression forms. Traditional seq2seq-based models on paraphrase generation mainly focus on the fidelity while ignoring the diversity of outputs. In this paper, we propose a deep generative model to generate diverse paraphrases. We build our model based on the conditional generative adversarial network, and propose to incorporate a simple yet effective diversity loss term into the model in order to improve the diversity of outputs. The proposed diversity loss maximizes the ratio of pairwise distance between the generated texts and their corresponding latent codes, forcing the generator to focus more on the latent codes and produce diverse samples. Experimental results on benchmarks of paraphrase generation show that our proposed model can generate more diverse paraphrases compared with baselines.
CITATION STYLE
Cao, Y., & Wan, X. (2020). DivGAN: Towards diverse paraphrase generation via diversified generative adversarial network. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 2411–2421). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.218
Mendeley helps you to discover research relevant for your work.