Reliability testing for natural language processing systems

15Citations
Citations of this article
94Readers
Mendeley users who have this article in their library.

Abstract

Questions of fairness, robustness, and transparency are paramount to address before deploying NLP systems. Central to these concerns is the question of reliability: Can NLP systems reliably treat different demographics fairly and function correctly in diverse and noisy environments? To address this, we argue for the need for reliability testing and contextualize it among existing work on improving accountability. We show how adversarial attacks can be reframed for this goal, via a framework for developing reliability tests. We argue that reliability testing - with an emphasis on interdisciplinary collaboration - will enable rigorous and targeted testing, and aid in the enactment and enforcement of industry standards.

Cite

CITATION STYLE

APA

Tan, S., Joty, S., Baxter, K., Taeihagh, A., Bennett, G. A., & Kan, M. Y. (2021). Reliability testing for natural language processing systems. In ACL-IJCNLP 2021 - 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Proceedings of the Conference (pp. 4153–4169). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.acl-long.321

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free