Data-driven approaches using neural networks have achieved promising performances in natural language generation (NLG). However, neural generators are prone to make mistakes, e.g., neglecting an input slot value and generating a redundant slot value. Prior works refer this to hallucination phenomenon. In this paper, we study slot consistency for building reliable NLG systems with all slot values of input dialogue act (DA) properly generated in output sentences. We propose Iterative Rectification Network (IRN) for improving general NLG systems to produce both correct and fluent responses. It applies a bootstrapping algorithm to sample training candidates and uses reinforcement learning to incorporate discrete reward related to slot inconsistency into training. Comprehensive studies have been conducted on multiple benchmark datasets, showing that the proposed methods have significantly reduced the slot error rate (ERR) for all strong baselines. Human evaluations also have confirmed its effectiveness.
CITATION STYLE
Li, Y., Yao, K., Qin, L., Che, W., Li, X., & Liu, T. (2020). Slot-consistent NLG for task-oriented dialogue systems with iterative rectification network. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 97–106). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.10
Mendeley helps you to discover research relevant for your work.