Abstract
Open domain question answering (ODQA) on a large-scale corpus of documents (e.g., Wikipedia) is a key challenge in computer science. Although transformer-based language models such as Bert have shown an ability to outperform humans to extract answers from small pre-selected passages of text, they suffer from their high complexity if the search space is much larger. The most common way to deal with this problem is to add a preliminary information retrieval step to strongly filter the corpus and keep only the relevant passages. In this article, the authors consider a more direct and complementary solution that consists of restricting the attention mechanism in transformer-based models to allow a more efficient management of computations. The resulting variants are competitive with the original models on the extractive task and allow, in the ODQA setting, a significant acceleration of predictions and sometimes even an improvement in the quality of response.
Author supplied keywords
Cite
CITATION STYLE
Siblini, W., Challal, M., & Pasqual, C. (2022). efficient open Domain Question Answering With Delayed Attention in Transformer-Based Models. International Journal of Data Warehousing and Mining, 18(2). https://doi.org/10.4018/IJDWM.298005
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.