Improving Syntactic Probing Correctness and Robustness with Control Tasks

1Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Syntactic probing methods have been used to examine whether and how pre-trained language models (PLMs) encode syntactic relations. However, the probing methods are usually biased by the PLMs’ memorization of common word co-occurrences, even if they do not form syntactic relations. This paper presents a random-word-substitution and random-label-matching control task to reduce these biases and improve the robustness of syntactic probing methods. Our control tasks are also shown to notably improve the consistency of probing results between different probing methods and make the methods more robust with respect to the text attributes of the probing instances. Our control tasks make syntactic probing methods better at reconstructing syntactic relations and more generalizable to unseen text domains. Our experiments show that our proposed control tasks are effective on different PLMs, probing methods, and syntactic relations.

Cite

CITATION STYLE

APA

Ma, W., Wang, B., Zhang, H., Wang, L., Coto-Solano, R., Hassanpour, S., & Vosoughi, S. (2023). Improving Syntactic Probing Correctness and Robustness with Control Tasks. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 2, pp. 402–415). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-short.35

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free