Paragraph-based Transformer Pre-training for Multi-Sentence Inference

8Citations
Citations of this article
47Readers
Mendeley users who have this article in their library.

Abstract

Inference tasks such as answer sentence selection (AS2) or fact verification are typically solved by fine-tuning transformer-based models as individual sentence-pair classifiers. Recent studies show that these tasks benefit from modeling dependencies across multiple candidate sentences jointly. In this paper, we first show that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks. We then propose a new pre-training objective that models the paragraph-level semantics across multiple input sentences. Our evaluation on three AS2 and one fact verification datasets demonstrates the superiority of our pre-training technique over the traditional ones for transformers used as joint models for multi-candidate inference tasks, as well as when used as cross-encoders for sentence-pair formulations of these tasks.

Cite

CITATION STYLE

APA

Di Liello, L., Garg, S., Soldaini, L., & Moschitti, A. (2022). Paragraph-based Transformer Pre-training for Multi-Sentence Inference. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 2521–2531). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.181

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free