efficient open Domain Question Answering With Delayed Attention in Transformer-Based Models

1Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Open domain question answering (ODQA) on a large-scale corpus of documents (e.g., Wikipedia) is a key challenge in computer science. Although transformer-based language models such as Bert have shown an ability to outperform humans to extract answers from small pre-selected passages of text, they suffer from their high complexity if the search space is much larger. The most common way to deal with this problem is to add a preliminary information retrieval step to strongly filter the corpus and keep only the relevant passages. In this article, the authors consider a more direct and complementary solution that consists of restricting the attention mechanism in transformer-based models to allow a more efficient management of computations. The resulting variants are competitive with the original models on the extractive task and allow, in the ODQA setting, a significant acceleration of predictions and sometimes even an improvement in the quality of response.

Cite

CITATION STYLE

APA

Siblini, W., Challal, M., & Pasqual, C. (2022). efficient open Domain Question Answering With Delayed Attention in Transformer-Based Models. International Journal of Data Warehousing and Mining, 18(2). https://doi.org/10.4018/IJDWM.298005

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free