Partial-input baselines show that NLI models can ignore context, but they don't

9Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.

Abstract

When strong partial-input baselines reveal artifacts in crowdsourced NLI datasets, the performance of full-input models trained on such datasets is often dismissed as reliance on spurious correlations. We investigate whether state-of-the-art NLI models are capable of overriding default inferences made by a partial-input baseline. We introduce an evaluation set of 600 examples consisting of perturbed premises to examine a RoBERTa model's sensitivity to edited contexts. Our results indicate that NLI models are still capable of learning to condition on context-a necessary component of inferential reasoning-despite being trained on artifact-ridden datasets.

Cite

CITATION STYLE

APA

Srikanth, N., & Rudinger, R. (2022). Partial-input baselines show that NLI models can ignore context, but they don’t. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 4753–4763). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.350

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free