Abstract
The popular skip-gram model induces word embeddings by exploiting the signal from word-context coocurrence. We offer a new interpretation of skip-gram based on exponential family PCA-a form of matrix factorization. This makes it clear that we can extend the skip-gram method to tensor factorization, in order to train embeddings through richer higher-order coocurrences, e.g., triples that include positional information (to incorporate syntax) or morphological information (to share parameters across related words). We experiment on 40 languages and show that our model improves upon skip-gram.
Cite
CITATION STYLE
Cotterell, R., Poliak, A., Van Durme, B., & Eisner, J. (2017). Explaining and generalizing skip-gram through exponential family principal component analysis. In 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017 - Proceedings of Conference (Vol. 2, pp. 175–181). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/e17-2028
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.