The generalization ability of pre-trained language models (PLMs) in downstream tasks is heavily influenced by fine-tuning. The objective of fine-tuning is to transform the latent representation of PLMs from a universal space to a target space, allowing the model to be applied to downstream tasks with the capability of generalizing to unseen samples. However, the effect of PLMs will be diminished when the training data coverage is insufficient, in which fine-tuning is inadequate to learn the complete mapping. In this study, we propose a new fine-tuning framework, referred to as G-Tuning, that aims to preserve the generalization ability of PLMs in downstream tasks. Specifically, we integrate a generative adversarial network into the fine-tuning process to aid in the transformation of the latent representation in the entire space. Empirical evaluations on the GLUE benchmark, as well as two additional demanding scenarios involving domain and language generalization, demonstrate that G-Tuning can accurately map the universal representation to the target space, thus effectively enhancing the generalization performance of PLMs across various downstream tasks.
CITATION STYLE
Weng, R., Cheng, W., & Zhang, M. (2023). G-Tuning: Improving Generalization of Pre-trained Language Models with Generative Adversarial Network. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 4747–4755). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.291
Mendeley helps you to discover research relevant for your work.