Transitioning from benchmarks to a real-world case of information-seeking in Scientific Publications

2Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Although recent years have been marked by incredible advances in the whole development process of NLP systems, there are still blind spots in characterizing what is still hampering real-world adoption of models in knowledge-intensive settings. In this paper, we illustrate through a real-world zero-shot text search case for information seeking in scientific papers, the masked phenomena that the current process of measuring performance might not reflect, even when benchmarks are, in appearance, faithfully representative of the task at hand. In addition to experimenting with TREC-COVID and NFCorpus, we provide an industrial, expert-carried/annotated, case of studying vitamin B's impact on health. We thus discuss the misalignment between solely focusing on single-metric performance as a criterion for model choice and relevancy as a subjective measure for meeting a user's need.

Cite

CITATION STYLE

APA

Tahri, C., Bochnakian, A., Haouat, P., & Tannier, X. (2023). Transitioning from benchmarks to a real-world case of information-seeking in Scientific Publications. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 1066–1076). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.68

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free