Climbing towards NLU: On meaning, form, and understanding in the age of data

915Citations
Citations of this article
1.0kReaders
Mendeley users who have this article in their library.

Abstract

The success of the large neural language models on many NLP tasks is exciting. However, we find that these successes sometimes lead to hype in which these models are being described as “understanding” language or capturing “meaning”. In this position paper, we argue that a system trained only on form has a priori no way to learn meaning. In keeping with the ACL 2020 theme of “Taking Stock of Where We've Been and Where We're Going”, we argue that a clear understanding of the distinction between form and meaning will help guide the field towards better science around natural language understanding.

Cite

CITATION STYLE

APA

Bender, E. M., & Koller, A. (2020). Climbing towards NLU: On meaning, form, and understanding in the age of data. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 5185–5198). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.463

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free