Semantic relations are core to how humans understand and express concepts in the real world using language. Recently, there has been a thread of research aimed at modeling these relations by learning vector representations from text corpora. Most of these approaches focus strictly on leveraging the co-occurrences of relationship word pairs within sentences. In this paper, we investigate the hypothesis that examples of a lexical relation in a corpus are fundamental to a neural word embedding’s ability to complete analogies involving the relation. Our experiments, in which we remove all known examples of a relation from training corpora, show only marginal degradation in analogy completion performance involving the removed relation. This finding enhances our understanding of neural word embeddings, showing that co-occurrence information of a particular semantic relation is the not the main source of their structural regularity.
CITATION STYLE
Chiang, H. Y., Camacho-Collados, J., & Pardos, Z. A. (2020). Understanding the Source of Semantic Regularities in Word Embeddings. In CoNLL 2020 - 24th Conference on Computational Natural Language Learning, Proceedings of the Conference (pp. 119–131). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.conll-1.9
Mendeley helps you to discover research relevant for your work.