Probing what different NLP tasks teach machines about function word comprehension

69Citations
Citations of this article
176Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We introduce a set of nine challenge tasks that test for the understanding of function words. These tasks are created by structurally mutating sentences from existing datasets to target the comprehension of specific types of function words (e.g., prepositions, wh-words). Using these probing tasks, we explore the effects of various pretraining objectives for sentence encoders (e.g., language modeling, CCG supertagging and natural language inference (NLI)) on the learned representations. Our results show that pretraining on language modeling performs the best on average across our probing tasks, supporting its widespread use for pretraining state-of-the-art NLP models, and CCG supertagging and NLI pretraining perform comparably. Overall, no pretraining objective dominates across the board, and our function word probing tasks highlight several intuitive differences between pretraining objectives, e.g., that NLI helps the comprehension of negation.

Cite

CITATION STYLE

APA

Kim, N., Patel, R., Poliak, A., Wang, A., Xia, P., McCoy, R. T., … Pavlick, E. (2019). Probing what different NLP tasks teach machines about function word comprehension. In *SEM@NAACL-HLT 2019 - 8th Joint Conference on Lexical and Computational Semantics (pp. 235–249). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/S19-1026

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free