Abstract
Previous studies have argued that pre-trained language models encode commonsense relational knowledge (e.g. that apples are edible). However, simultaneous work has revealed that such models are often insensitive to context, even ignoring overt contextual cues such as negations. In this paper, we investigate whether masked language models (the BERT family) can move beyond naive associative biases (e.g., apple ? edible) when the context warrants (e.g. ranking inedible higher when presented with the information that the apple is rotten). We introduce the WINOVENTI procedure, which adversarially exploits generic associations in masked language models to create model-specific Winograd-style entailment schemas. Using our constructed WINOVENTI challenges set of over 2, 000 schemas, we show that language models in the BERT family experience a steep drop in performance on prompts that require them to pick answers which require reasoning about context (e.g., from 89.8% to 18.4% for BERTLARGE). We present evidence that language models exhibit different associative biases, suggesting a need for future work in developing and analyzing frameworks similar to WINOVENTI that are tuned to model-specific weaknesses.
Cite
CITATION STYLE
Do, N., & Pavlick, E. (2021). Are Rotten Apples Edible? Challenging Commonsense Inference Ability with Exceptions. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 (pp. 2061–2073). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-acl.181
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.