GENSF: Simultaneous Adaptation of Generative Pre-trained Models and Slot Filling

9Citations
Citations of this article
54Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In transfer learning, it is imperative to achieve strong alignment between a pre-trained model and a downstream task. Prior work has done this by proposing task-specific pre-training objectives, which sacrifices the inherent scalability of the transfer learning paradigm. We instead achieve strong alignment by simultaneously modifying both the pre-trained model and the formulation of the downstream task, which is more efficient and preserves the scalability of transfer learning. We present GENSF (Generative Slot Filling), which leverages a generative pre-trained open-domain dialog model for slot filling. GENSF (1) adapts the pre-trained model by incorporating inductive biases about the task and (2) adapts the downstream task by reformulating slot filling to better leverage the pre-trained model's capabilities. GENSF achieves state-of-the-art results on two slot filling datasets with strong gains in few-shot and zero-shot settings. We achieve a 9 F1 score improvement in zeroshot slot filling. This highlights the value of strong alignment between the pre-trained model and the downstream task.

Cite

CITATION STYLE

APA

Mehri, S., & Eskenazi, M. (2021). GENSF: Simultaneous Adaptation of Generative Pre-trained Models and Slot Filling. In SIGDIAL 2021 - 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue, Proceedings of the Conference (pp. 489–498). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.sigdial-1.51

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free