Multimodal word distributions

59Citations
Citations of this article
407Readers
Mendeley users who have this article in their library.

Abstract

Word embeddings provide point representations of words containing useful semantic information. We introduce multimodal word distributions formed from Gaussian mixtures, for multiple word meanings, entailment, and rich uncertainty information. To learn these distributions, we propose an energy-based max-margin objective. We show that the resulting approach captures uniquely expressive semantic information, and outperforms alternatives, such as word2vec skip-grams, and Gaussian embeddings, on benchmark datasets such as word similarity and entailment.

Cite

CITATION STYLE

APA

Athiwaratkun, B., & Wilson, A. G. (2017). Multimodal word distributions. In ACL 2017 - 55th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 1, pp. 1645–1656). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/P17-1151

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free