Entity-Conditioned Question Generation for Robust Attention Distribution in Neural Information Retrieval

3Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We show that supervised neural information retrieval (IR) models are prone to learning sparse attention patterns over passage tokens, which can result in key phrases including named entities receiving low attention weights, eventually leading to model under-performance. Using a novel targeted synthetic data generation method that identifies poorly attended entities and conditions the generation episodes on those, we teach neural IR to attend more uniformly and robustly to all entities in a given passage. On two public IR benchmarks, we empirically show that the proposed method helps improve both the model's attention patterns and retrieval performance, including in zero-shot settings.

Cite

CITATION STYLE

APA

Reddy, R. G., Sultan, M. A., Franz, M., Sil, A., & Ji, H. (2022). Entity-Conditioned Question Generation for Robust Attention Distribution in Neural Information Retrieval. In SIGIR 2022 - Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (pp. 2462–2466). Association for Computing Machinery, Inc. https://doi.org/10.1145/3477495.3531878

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free