Abstract
This paper introduces a simple yet effective data-centric approach for the task of improving persona-conditioned dialogue agents. Prior model-centric approaches unquestioningly depend on the raw crowdsourced benchmark datasets such as Persona-Chat. In contrast, we aim to fix annotation artifacts in benchmarking, which is orthogonally applicable to any dialogue model. Specifically, we augment relevant personas to improve dialogue dataset/agent, by leveraging the primal-dual structure of the two tasks, predicting dialogue responses and personas based on each other. Experiments on Persona-Chat show that our approach outperforms pretrained LMs by an 11.7 point gain in terms of accuracy.
Cite
CITATION STYLE
Kim, M., Kwak, B. W., Kim, Y., Lee, H. I., Hwang, S. W., & Yeo, J. (2022). Dual Task Framework for Improving Persona-Grounded Dialogue Dataset. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 10912–10920). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i10.21338
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.