Unsupervised Representation Disentanglement of Text: An Evaluation on Synthetic Datasets

2Citations
Citations of this article
45Readers
Mendeley users who have this article in their library.

Abstract

To highlight the challenges of achieving representation disentanglement for text domain in an unsupervised setting, in this paper we select a representative set of successfully applied models from the image domain. We evaluate these models on 6 disentanglement metrics, as well as on downstream classification tasks and homotopy. To facilitate the evaluation, we propose two synthetic datasets with known generative factors. Our experiments highlight the existing gap in the text domain and illustrate that certain elements such as representation sparsity (as an inductive bias), or representation coupling with the decoder could impact disentanglement. To the best of our knowledge, our work is the first attempt on the intersection of unsupervised representation disentanglement and text, and provides the experimental framework and datasets for examining future developments in this direction.

Cite

CITATION STYLE

APA

Zhang, L., Prokhorov, V., & Shareghi, E. (2021). Unsupervised Representation Disentanglement of Text: An Evaluation on Synthetic Datasets. In RepL4NLP 2021 - 6th Workshop on Representation Learning for NLP, Proceedings of the Workshop (pp. 128–140). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.repl4nlp-1.14

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free