Word associations and the distance properties of context-aware word embeddings

7Citations
Citations of this article
67Readers
Mendeley users who have this article in their library.

Abstract

What do people know when they know the meaning of words? Word associations have been widely used to tap into lexical representations and their structure, as a way of probing semantic knowledge in humans. We investigate whether current word embedding spaces (contextualized and uncontextualized) can be considered good models of human lexical knowledge by studying whether they have comparable characteristics to human association spaces. We study the three properties of association rank, asymmetry of similarity and triangle inequality. We find that word embeddings are good models of some word associations properties. They replicate well human associations between words, and, like humans, their context-aware variants show violations of the triangle inequality. While they do show asymmetry of similarities, their asymmetries do not map those of human association norms.

References Powered by Scopus

WordNet: A Lexical Database for English

11660Citations
N/AReaders
Get full text

Features of similarity

5457Citations
N/AReaders
Get full text

A Solution to Plato's Problem: The Latent Semantic Analysis Theory of Acquisition, Induction, and Representation of Knowledge

4571Citations
N/AReaders
Get full text

Cited by Powered by Scopus

The World of an Octopus: How Reporting Bias Influences a Language Model's Perception of Color

22Citations
N/AReaders
Get full text

Beyond the Benchmarks: Toward Human-Like Lexical Representations

6Citations
N/AReaders
Get full text

A study on surprisal and semantic relatedness for eye-tracking data prediction

4Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Rodriguez, M. A., & Merlo, P. (2020). Word associations and the distance properties of context-aware word embeddings. In CoNLL 2020 - 24th Conference on Computational Natural Language Learning, Proceedings of the Conference (pp. 376–385). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.conll-1.30

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 21

72%

Researcher 6

21%

Lecturer / Post doc 2

7%

Readers' Discipline

Tooltip

Computer Science 25

74%

Linguistics 6

18%

Engineering 2

6%

Neuroscience 1

3%

Save time finding and organizing research with Mendeley

Sign up for free