Abstract
IR-based Question Answering (QA) systems typically use a sentence selector to extract the answer from retrieved documents. Recent studies have shown that powerful neural models based on the Transformer can provide an accurate solution to Answer Sentence Selection (AS2). Unfortunately, their computation cost prevents their use in real-world applications. In this paper, we show that standard and efficient neural rerankers can be used to reduce the amount of sentence candidates fed to Transformer models without hurting Accuracy, thus improving efficiency up to four times. This is an important finding as the internal representation of shallower neural models is dramatically different from the one used by a Transformer model, e.g., word vs. contextual embeddings.
Author supplied keywords
Cite
CITATION STYLE
Matsubara, Y., Vu, T., & Moschitti, A. (2020). Reranking for Efficient Transformer-based Answer Selection. In SIGIR 2020 - Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval (pp. 1577–1580). Association for Computing Machinery, Inc. https://doi.org/10.1145/3397271.3401266
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.