Abstract
Inference tasks such as answer sentence selection (AS2) or fact verification are typically solved by fine-tuning transformer-based models as individual sentence-pair classifiers. Recent studies show that these tasks benefit from modeling dependencies across multiple candidate sentences jointly. In this paper, we first show that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks. We then propose a new pre-training objective that models the paragraph-level semantics across multiple input sentences. Our evaluation on three AS2 and one fact verification datasets demonstrates the superiority of our pre-training technique over the traditional ones for transformers used as joint models for multi-candidate inference tasks, as well as when used as cross-encoders for sentence-pair formulations of these tasks.
Cite
CITATION STYLE
Di Liello, L., Garg, S., Soldaini, L., & Moschitti, A. (2022). Paragraph-based Transformer Pre-training for Multi-Sentence Inference. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 2521–2531). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.181
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.