AdapterShare: Task Correlation Modeling with Adapter Differentiation

0Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.

Abstract

Thanks to the development of pre-trained language models, multitask learning (MTL) methods have achieved great success in natural language understanding. However, current MTL methods pay more attention to task selection or model design to fuse as much knowledge as possible, while the intrinsic task correlation is often neglected. It is important to learn sharing strategies among multiple tasks rather than sharing everything. In this paper, we propose AdapterShare, an adapter differentiation method to explicitly model task correlation among multiple tasks. AdapterShare is automatically learned based on the gradients on tiny held-out validation data. Compared to single-task learning and fully shared MTL methods, our proposed method obtains obvious performance improvements. Compared to the existing MTL method AdapterFusion, AdapterShare achieves an absolute average improvement of 1.90 points on five dialogue understanding tasks and 2.33 points on NLU tasks. Our implementation is available at https://github.com/microsoft/ContextualSP.

Cite

CITATION STYLE

APA

Chen, Z., Chen, B., Chen, L., Yu, K., & Lou, J. G. (2022). AdapterShare: Task Correlation Modeling with Adapter Differentiation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 10645–10651). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.728

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free