G-Tuning: Improving Generalization of Pre-trained Language Models with Generative Adversarial Network

0Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.

Abstract

The generalization ability of pre-trained language models (PLMs) in downstream tasks is heavily influenced by fine-tuning. The objective of fine-tuning is to transform the latent representation of PLMs from a universal space to a target space, allowing the model to be applied to downstream tasks with the capability of generalizing to unseen samples. However, the effect of PLMs will be diminished when the training data coverage is insufficient, in which fine-tuning is inadequate to learn the complete mapping. In this study, we propose a new fine-tuning framework, referred to as G-Tuning, that aims to preserve the generalization ability of PLMs in downstream tasks. Specifically, we integrate a generative adversarial network into the fine-tuning process to aid in the transformation of the latent representation in the entire space. Empirical evaluations on the GLUE benchmark, as well as two additional demanding scenarios involving domain and language generalization, demonstrate that G-Tuning can accurately map the universal representation to the target space, thus effectively enhancing the generalization performance of PLMs across various downstream tasks.

Cite

CITATION STYLE

APA

Weng, R., Cheng, W., & Zhang, M. (2023). G-Tuning: Improving Generalization of Pre-trained Language Models with Generative Adversarial Network. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 4747–4755). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.291

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free