Understanding the Source of Semantic Regularities in Word Embeddings

9Citations
Citations of this article
66Readers
Mendeley users who have this article in their library.

Abstract

Semantic relations are core to how humans understand and express concepts in the real world using language. Recently, there has been a thread of research aimed at modeling these relations by learning vector representations from text corpora. Most of these approaches focus strictly on leveraging the co-occurrences of relationship word pairs within sentences. In this paper, we investigate the hypothesis that examples of a lexical relation in a corpus are fundamental to a neural word embedding’s ability to complete analogies involving the relation. Our experiments, in which we remove all known examples of a relation from training corpora, show only marginal degradation in analogy completion performance involving the removed relation. This finding enhances our understanding of neural word embeddings, showing that co-occurrence information of a particular semantic relation is the not the main source of their structural regularity.

Cite

CITATION STYLE

APA

Chiang, H. Y., Camacho-Collados, J., & Pardos, Z. A. (2020). Understanding the Source of Semantic Regularities in Word Embeddings. In CoNLL 2020 - 24th Conference on Computational Natural Language Learning, Proceedings of the Conference (pp. 119–131). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.conll-1.9

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free