Scientific and Creative Analogies in Pretrained Language Models

11Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

This paper examines the encoding of analogy in large-scale pretrained language models, such as BERT and GPT-2. Existing analogy datasets typically focus on a limited set of analogical relations, with a high similarity of the two domains between which the analogy holds. As a more realistic setup, we introduce the Scientific and Creative Analogy dataset (SCAN), a novel analogy dataset containing systematic mappings of multiple attributes and relational structures across dissimilar domains. Using this dataset, we test the analogical reasoning capabilities of several widely-used pretrained language models (LMs). We find that state-of-the-art LMs achieve low performance on these complex analogy tasks, highlighting the challenges still posed by analogy understanding.

Cite

CITATION STYLE

APA

Czinczoll, T., Yannakoudakis, H., Mishra, P., & Shutova, E. (2022). Scientific and Creative Analogies in Pretrained Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 2094–2100). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.153

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free