Word representation on small background texts

1Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Vector representations of words learned from large scale background texts can be used as useful features in natural language processing and machine learning applications. Word representations in previous works were often trained on large-scale unlabeled texts. However, in some scenarios, large scale background texts are not available. Therefore, in this paper, we propose a novel word representation model based on maximum-margin to train word representation using small set of background texts. Experimental results show many advantages of our method.

Cite

CITATION STYLE

APA

Li, L., Jiang, Z., Liu, Y., & Huang, D. (2016). Word representation on small background texts. In Communications in Computer and Information Science (Vol. 669, pp. 143–150). Springer Verlag. https://doi.org/10.1007/978-981-10-2993-6_12

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free