Abstract
We present a new method for estimating vector space representations of words: embedding learning by concept induction. We test this method on a highly parallel corpus and learn semantic representations of words in 1259 different languages in a single common space. An extensive experimental evaluation on crosslingual word similarity and sentiment analysis indicates that concept-based multilingual embedding learning performs better than previous approaches.
Cite
CITATION STYLE
Dufter, P., Zhao, M., Schmitt, M., Fraser, A., & Schütze, H. (2018). Embedding learning through multilingual concept induction. In ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 1, pp. 1520–1530). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p18-1141
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.