Evaluating natural alpha embeddings on intrinsic and extrinsic tasks

3Citations
Citations of this article
61Readers
Mendeley users who have this article in their library.

Abstract

Skip-Gram is a simple, but effective, model to learn a word embedding mapping by estimating a conditional probability distribution for each word of the dictionary. In the context of Information Geometry, these distributions form a Riemannian statistical manifold, where word embeddings are interpreted as vectors in the tangent bundle of the manifold. In this paper we show how the choice of the geometry on the manifold allows impacts on the performances both on intrinsic and extrinsic tasks, in function of a deformation parameter alpha.

Cite

CITATION STYLE

APA

Volpi, R., & Malagò, L. (2020). Evaluating natural alpha embeddings on intrinsic and extrinsic tasks. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 61–71). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.repl4nlp-1.9

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free