Large-scale Lifelong Learning of In-context Instructions and How to Tackle It

3Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

Abstract

Jointly fine-tuning a Pre-trained Language Model (PLM) on a pre-defined set of tasks with in-context instructions has been proven to improve its generalization performance, allowing us to build a universal language model that can be deployed across task boundaries. In this work, we explore for the first time whether this attractive property of in-context instruction learning can be extended to a scenario in which tasks are fed to the target PLM in a sequential manner. The primary objective of so-called lifelong in-context instruction learning is to improve the target PLM's instance- and task-level generalization performance as it observes more tasks. DYNAINST, the proposed method to lifelong in-context instruction learning, achieves noticeable improvements in both types of generalization, nearly reaching the upper bound performance obtained through joint training.

Cite

CITATION STYLE

APA

Mok, J., Do, J., Lee, S., Taghavi, T., Yu, S., & Yoon, S. (2023). Large-scale Lifelong Learning of In-context Instructions and How to Tackle It. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 12573–12589). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.703

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free