Prompt Learning for Domain Adaptation in Task-Oriented Dialogue

2Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

Conversation designers continue to face significant obstacles when creating production-quality task-oriented dialogue systems. The complexity and cost involved in schema development and data collection is often a major barrier for such designers, limiting their ability to create natural, user-friendly experiences. We frame the classification of user intent as the generation of a canonical form, a lightweight semantic representation using natural language. We show that canonical forms offer a promising alternative to traditional methods for intent classification. By tuning soft prompts for a frozen large language model, we show that canonical forms generalize very well to new, unseen domains in a zero- or few-shot setting. The method is also sample-efficient, reducing the complexity and effort of developing new task-oriented dialogue domains.

Cite

CITATION STYLE

APA

Sreedhar, M. N., & Parisien, C. (2022). Prompt Learning for Domain Adaptation in Task-Oriented Dialogue. In SereTOD 2022 - Towards Semi-Supervised and Reinforced Task-Oriented Dialog Systems, Proceedings of the Workshop (pp. 24–30). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.seretod-1.4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free