Abstract
We present a method for applying a neural network trained on one (resource-rich) language for a given task to other (resource-poor) languages. We accomplish this by inducing a mapping from pre-trained cross-lingual word embeddings to the embedding layer of the neural network trained on the resource-rich language. To perform element-wise cross-task embedding projection, we invent locally linear mapping which assumes and preserves the local topology across the semantic spaces before and after the projection. Experimental results on topic classification task and sentiment analysis task showed that the fully task-specific multilingual model obtained using our method outperformed the existing multilingual models with embedding layers fixed to pre-trained cross-lingual word embeddings.
Cite
CITATION STYLE
Sakuma, J., & Yoshinaga, N. (2019). Multilingual model using cross-task embedding projection. In CoNLL 2019 - 23rd Conference on Computational Natural Language Learning, Proceedings of the Conference (pp. 22–32). Association for Computational Linguistics. https://doi.org/10.18653/v1/k19-1003
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.