Jointly fine-tuning a Pre-trained Language Model (PLM) on a pre-defined set of tasks with in-context instructions has been proven to improve its generalization performance, allowing us to build a universal language model that can be deployed across task boundaries. In this work, we explore for the first time whether this attractive property of in-context instruction learning can be extended to a scenario in which tasks are fed to the target PLM in a sequential manner. The primary objective of so-called lifelong in-context instruction learning is to improve the target PLM's instance- and task-level generalization performance as it observes more tasks. DYNAINST, the proposed method to lifelong in-context instruction learning, achieves noticeable improvements in both types of generalization, nearly reaching the upper bound performance obtained through joint training.
CITATION STYLE
Mok, J., Do, J., Lee, S., Taghavi, T., Yu, S., & Yoon, S. (2023). Large-scale Lifelong Learning of In-context Instructions and How to Tackle It. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 12573–12589). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.703
Mendeley helps you to discover research relevant for your work.