Reranking for Efficient Transformer-based Answer Selection

41Citations
Citations of this article
25Readers
Mendeley users who have this article in their library.
Get full text

Abstract

IR-based Question Answering (QA) systems typically use a sentence selector to extract the answer from retrieved documents. Recent studies have shown that powerful neural models based on the Transformer can provide an accurate solution to Answer Sentence Selection (AS2). Unfortunately, their computation cost prevents their use in real-world applications. In this paper, we show that standard and efficient neural rerankers can be used to reduce the amount of sentence candidates fed to Transformer models without hurting Accuracy, thus improving efficiency up to four times. This is an important finding as the internal representation of shallower neural models is dramatically different from the one used by a Transformer model, e.g., word vs. contextual embeddings.

Cite

CITATION STYLE

APA

Matsubara, Y., Vu, T., & Moschitti, A. (2020). Reranking for Efficient Transformer-based Answer Selection. In SIGIR 2020 - Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval (pp. 1577–1580). Association for Computing Machinery, Inc. https://doi.org/10.1145/3397271.3401266

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free