A Universal Discriminator for Zero-Shot Generalization

7Citations
Citations of this article
22Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Generative modeling has been the dominant approach for large-scale pretraining and zero-shot generalization. In this work, we challenge this convention by showing that discriminative approaches perform substantially better than generative ones on a large number of NLP tasks. Technically, we train a single discriminator to predict whether a text sample comes from the true data distribution, similar to GANs. Since many NLP tasks can be formulated as selecting from a few options, we use this discriminator to predict the concatenation of input and which option has the highest probability of coming from the true data distribution. This simple formulation achieves state-of-the-art zero-shot results on the T0 benchmark, outperforming T0 by 16.0%, 7.8%, and 11.5% respectively on different scales. In the finetuning setting, our approach also achieves new state-of-the-art results on a wide range of NLP tasks, with only 1/4 parameters of previous methods. Meanwhile, our approach requires minimal prompting efforts, which largely improves robustness and is essential for real-world applications. Furthermore, we also jointly train a generalized UD in combination with generative tasks, which maintains its advantage on discriminative tasks and simultaneously works on generative tasks.

Cite

CITATION STYLE

APA

Xu, H., Lin, Z., Zhou, J., Zheng, Y., & Yang, Z. (2023). A Universal Discriminator for Zero-Shot Generalization. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 10559–10575). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.589

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free