MULTIINSTRUCT: Improving Multi-Modal Zero-Shot Learning via Instruction Tuning

38Citations
Citations of this article
83Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Instruction tuning, a new learning paradigm that fine-tunes pre-trained language models on tasks specified through instructions, has shown promising zero-shot performance on various natural language processing tasks. However, it has yet to be explored for vision and multimodal tasks. In this work, we introduce MULTIINSTRUCT, the first multimodal instruction tuning benchmark dataset that consists of 62 diverse multimodal tasks in a unified seq-to-seq format covering 10 broad categories. The tasks are derived from 21 existing open-source datasets and each task is equipped with 5 expert-written instructions. We take OFA (Wang et al., 2022a) as the base pre-trained model for multimodal instruction tuning, and to further improve its zero-shot performance, we explore multiple transfer learning strategies to leverage the large-scale NATURAL INSTRUCTIONS dataset (Mishra et al., 2022). Experimental results demonstrate strong zero-shot performance on various unseen multimodal tasks and the benefit of transfer learning from a text-only instruction dataset. We also design a new evaluation metric - Sensitivity, to evaluate how sensitive the model is to the variety of instructions. Our results indicate that fine-tuning the model on a diverse set of tasks and instructions leads to a reduced sensitivity to variations in instructions for each task.

Cite

CITATION STYLE

APA

Xu, Z., Shen, Y., & Huang, L. (2023). MULTIINSTRUCT: Improving Multi-Modal Zero-Shot Learning via Instruction Tuning. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 11445–11465). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.641

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free