ZeroAE: Pre-trained Language Model based Autoencoder for Transductive Zero-shot Text Classification

4Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Many text classification tasks require handling unseen domains with plenty of unlabeled data, thus giving rise to the self-adaption or the so-called transductive zero-shot learning (TZSL) problem. However, current methods based solely on encoders or decoders overlook the possibility that these two modules may promote each other. As a first effort to bridge this gap, we propose an autoencoder named ZeroAE. Specifically, the text is encoded with two separate BERT-based encoders into two disentangled spaces, i.e., label-relevant (for classification) and label-irrelevant respectively. The two latent spaces are then decoded by prompting GPT-2 to recover the text as well as to further generate text with labels in the unseen domains to train the encoder in turn. To better exploit the unlabeled data, a novel indirect uncertainty-aware sampling (IUAS) approach is proposed to train ZeroAE. Extensive experiments show that ZeroAE largely surpasses the SOTA methods by 15.93% and 8.70% on average respectively in the label-partially-unseen and label-fully-unseen scenario. Notably, the label-fully-unseen ZeroAE even possesses superior performance to the label-partially-unseen SOTA methods.

Cite

CITATION STYLE

APA

Guo, K., Yu, H., Liao, C., Li, J., & Zhang, H. (2023). ZeroAE: Pre-trained Language Model based Autoencoder for Transductive Zero-shot Text Classification. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 3202–3219). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.200

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free