A gloss composition and context clustering based distributed word sense representation model

8Citations
Citations of this article
28Readers
Mendeley users who have this article in their library.

Abstract

In recent years, there has been an increasing interest in learning a distributed representation of word sense. Traditional context clustering based models usually require careful tuning of model parameters, and typically perform worse on infrequent word senses. This paper presents a novel approach which addresses these limitations by first initializing the word sense embeddings through learning sentence-level embeddings from WordNet glosses using a convolutional neural networks. The initialized word sense embeddings are used by a context clustering based model to generate the distributed representations of word senses. Our learned representations outperform the publicly available embeddings on half of the metrics in the word similarity task, 6 out of 13 sub tasks in the analogical reasoning task, and gives the best overall accuracy in the word sense effect classification task, which shows the effectiveness of our proposed distributed distribution learning model.

Cite

CITATION STYLE

APA

Chen, T., Xu, R., He, Y., & Wang, X. (2015). A gloss composition and context clustering based distributed word sense representation model. Entropy, 17(9), 6007–6024. https://doi.org/10.3390/e17096007

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free