Many Languages, One Parser

  • Ammar W
  • Mulcaire G
  • Ballesteros M
  • et al.
N/ACitations
Citations of this article
197Readers
Mendeley users who have this article in their library.

Abstract

We train one multilingual model for dependency parsing and use it to parse sentences in several languages. The parsing model uses (i) multilingual word clusters and embeddings; (ii) token-level language information; and (iii) language-specific features (fine-grained POS tags). This input representation enables the parser not only to parse effectively in multiple languages, but also to generalize across languages based on linguistic universals and typological similarities, making it more effective to learn from limited annotations. Our parser’s performance compares favorably to strong baselines in a range of data scenarios, including when the target language has a large treebank, a small treebank, or no treebank for training.

Cite

CITATION STYLE

APA

Ammar, W., Mulcaire, G., Ballesteros, M., Dyer, C., & Smith, N. A. (2016). Many Languages, One Parser. Transactions of the Association for Computational Linguistics, 4, 431–444. https://doi.org/10.1162/tacl_a_00109

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free