Context vectors are reflections of word vectors in half the dimensions

ISSN: 10450823
0Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.

Abstract

This paper takes a step towards the theoretical analysis of the relationship between word embeddings and context embeddings in models such as word2vec. We start from basic probabilistic assumptions on the nature of word vectors, context vectors, and text generation. These assumptions are supported either empirically or theoretically by the existing literature. Next, we show that under these assumptions the widely-used word-word PMI matrix is approximately a random symmetric Gaussian ensemble. This, in turn, implies that context vectors are reflections of word vectors in approximately half the dimensions. As a direct application of our result, we suggest a theoretically grounded way of tying weights in the SGNS model.

Cite

CITATION STYLE

APA

Assylbekov, Z., & Takhanov, R. (2020). Context vectors are reflections of word vectors in half the dimensions. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 5115–5119). International Joint Conferences on Artificial Intelligence.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free