Cross-lingual syntactically informed distributed word representations

10Citations
Citations of this article
86Readers
Mendeley users who have this article in their library.

Abstract

We develop a novel cross-lingual word representation model which injects syntactic information through dependencybased contexts into a shared cross-lingual word vector space. The model, termed CLDEPEMB, is based on the following assumptions: (1) dependency relations are largely language-independent, at least for related languages and prominent dependency links such as direct objects, as evidenced by the Universal Dependencies project; (2) word translation equivalents take similar grammatical roles in a sentence and are therefore substitutable within their syntactic contexts. Experiments with several language pairs on word similarity and bilingual lexicon induction, two fundamental semantic tasks emphasising semantic similarity, suggest the usefulness of the proposed syntactically informed crosslingual word vector spaces. Improvements are observed in both tasks over standard cross-lingual "offline mapping" baselines trained using the same setup and an equal level of bilingual supervision.

Cite

CITATION STYLE

APA

Vulic, I. (2017). Cross-lingual syntactically informed distributed word representations. In 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017 - Proceedings of Conference (Vol. 2, pp. 408–414). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/e17-2065

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free