Abstract
The alignment of word embedding spaces in different languages into a common crosslingual space has recently been in vogue. Strategies that do so compute pairwise alignments and then map multiple languages to a single pivot language (most often English). These strategies, however, are biased towards the choice of the pivot language, given that language proximity and the linguistic characteristics of the target language can strongly impact the resultant crosslingual space in detriment of topologically distant languages. We present a strategy that eliminates the need for a pivot language by learning the mappings across languages in a hierarchical way. Experiments demonstrate that our strategy significantly improves vocabulary induction scores in all existing benchmarks, as well as in a new non-English–centered benchmark we built, which we make publicly available.
Cite
CITATION STYLE
Azpiazu, I. M., & Pera, M. S. (2020). Hierarchical mapping for crosslingual word embedding alignment. Transactions of the Association for Computational Linguistics, 8, 361–376. https://doi.org/10.1162/tacl_a_00320
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.