In this paper, we study the utilization of pretrained language models to enable few-shot Natural Language Generation (NLG) in taskoriented dialog systems. We introduce a system consisting of iterative self-training and an extensible mini-template framework that textualizes the structured input data into seminatural text to fully take advantage of pretrained language models. We compare various representations of NLG models' input and output and show that transforming the input and output to be similar to what the language model has seen before during pretraining improves the model's few-shot performance substantially. We show that neural models can be trained with as few as 300 annotated examples while providing high fidelity, considerably lowering the resource requirements for standing up a new domain or language. This level of data efficiency removes the need for crowd-sourced data collection resulting in higher quality data annotated by expert linguists. In addition, model maintenance and debugging processes will improve in this fewshot setting. Finally, we explore distillation and using a caching system to satisfy latency requirements of real-world systems.
CITATION STYLE
Heidari, P., Einolghozati, A., Jain, S., Batra, S., Callender, L., Arun, A., … White, M. (2021). Getting to Production with Few-shot Natural Language Generation Models. In SIGDIAL 2021 - 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue, Proceedings of the Conference (pp. 66–76). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.sigdial-1.8
Mendeley helps you to discover research relevant for your work.