Abstract
In a multi-turn knowledge-grounded dialog, the difference between the knowledge selected at different turns usually provides potential clues to knowledge selection, which has been largely neglected in previous research. In this paper, we propose a difference-aware knowledge selection method. It first computes the difference between the candidate knowledge sentences provided at the current turn and those chosen in the previous turns. Then, the differential information is fused with or disentangled from the contextual information to facilitate final knowledge selection. Automatic, human observational, and interactive evaluation shows that our method is able to select knowledge more accurately and generate more informative responses, significantly outperforming the state-of-the-art baselines. The codes are available at https://github.com/chujiezheng/DiffKS.
Cite
CITATION STYLE
Zheng, C., Cao, Y., Jiang, D., & Huang, M. (2020). Difference-aware knowledge selection for knowledge-grounded conversation generation. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 115–125). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.11
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.