Evidentiality-guided Generation for Knowledge-Intensive NLP Tasks

19Citations
Citations of this article
73Readers
Mendeley users who have this article in their library.

Abstract

Retrieval-augmented generation models have shown state-of-the-art performance across many knowledge-intensive NLP tasks such as open-domain question answering and fact verification. These models are trained to generate a final output given retrieved passages that can be irrelevant to an input query, leading to learning spurious cues or memorization. This work introduces a method to incorporate evidentiality of passages-whether a passage contains correct evidence to support the output-into training the generator. We introduce a multi-task learning framework to jointly generate the final output and predict the evidentiality of each passage. Furthermore, we introduce a new task-agnostic method for obtaining high-quality silver evidentiality labels, addressing the issues of gold evidentiality labels being unavailable in most domains. Our experiments on five datasets across three knowledge-intensive tasks show that our new evidentiality-guided generator significantly outperforms its direct counterpart on all of them, and advances the state of the art on three of them. Our analysis shows that the multi-task learning and silver evidentiality mining play key roles.

Cite

CITATION STYLE

APA

Asai, A., Gardner, M., & Hajishirzi, H. (2022). Evidentiality-guided Generation for Knowledge-Intensive NLP Tasks. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 2226–2243). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.162

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free