CoDA21: Evaluating Language Understanding Capabilities of NLP Models With Context-Definition Alignment

2Citations
Citations of this article
39Readers
Mendeley users who have this article in their library.

Abstract

Pretrained language models (PLMs) have achieved superhuman performance on many benchmarks, creating a need for harder tasks. We introduce CoDA21 (Context Definition Alignment), a challenging benchmark that measures natural language understanding (NLU) capabilities of PLMs: Given a definition and a context each for k words, but not the words themselves, the task is to align the k definitions with the k contexts. CoDA21 requires a deep understanding of contexts and definitions, including complex inference and world knowledge. We find that there is a large gap between human and PLM performance, suggesting that CoDA21 measures an aspect of NLU that is not sufficiently covered in existing benchmarks.

Cite

CITATION STYLE

APA

Senel, L. K., Schick, T., & Schütze, H. (2022). CoDA21: Evaluating Language Understanding Capabilities of NLP Models With Context-Definition Alignment. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 2, pp. 815–824). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-short.92

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free