CESAR: Automatic Induction of Compositional Instructions for Multi-turn Dialogs

0Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

Instruction-based multitasking has played a critical role in the success of large language models (LLMs) in multi-turn dialog applications. While publicly-available LLMs have shown promising performance, when exposed to complex instructions with multiple constraints, they lag against state-of-the-art models like ChatGPT. In this work, we hypothesize that the availability of large-scale complex demonstrations is crucial in bridging this gap. Focusing on dialog applications, we propose a novel framework, CESAR, that unifies a large number of dialog tasks in the same format and allows programmatic induction of complex instructions without any manual effort. We apply CESAR on InstructDial, a benchmark for instruction-based dialog tasks. We further enhance InstructDial with new datasets and tasks and utilize CESAR to induce complex tasks with compositional instructions. This results in a new benchmark called InstructDial++, which includes 63 datasets with 86 basic tasks and 68 composite tasks. Through rigorous experiments, we demonstrate the scalability of CESAR in providing rich instructions. Models trained on InstructDial++ can follow compositional prompts, such as prompts that ask for multiple stylistic constraints.

Cite

CITATION STYLE

APA

Aksu, T., Hazarika, D., Mehri, S., Kim, S., Hakkani-Tür, D., Liu, Y., & Namazifar, M. (2023). CESAR: Automatic Induction of Compositional Instructions for Multi-turn Dialogs. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 11709–11737). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.717

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free