On the data requirements of probing

3Citations
Citations of this article
33Readers
Mendeley users who have this article in their library.

Abstract

As large and powerful neural language models are developed, researchers have been increasingly interested in developing diagnostic tools to probe them. There are many papers with conclusions of the form “observation X is found in model Y ”, using their own datasets with varying sizes. Larger probing datasets bring more reliability, but are also expensive to collect. There is yet to be a quantitative method for estimating reasonable probing dataset sizes. We tackle this omission in the context of comparing two probing configurations: after we have collected a small dataset from a pilot study, how many additional data samples are sufficient to distinguish two different configurations? We present a novel method to estimate the required number of data samples in such experiments and, across several case studies, we verify that our estimations have sufficient statistical power. Our framework helps to systematically construct probing datasets to diagnose neural NLP models.

Cite

CITATION STYLE

APA

Zhu, Z., Wang, J., Li, B., & Rudzicz, F. (2022). On the data requirements of probing. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 4132–4147). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-acl.326

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free