Improving Low-resource RRG Parsing with Structured Gloss Embeddings

0Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Treebanking for local languages is hampered by the lack of existing parsers to generate pre-annotations. However, it has been shown that reasonably accurate parsers can be bootstrapped with little initial training data when use is made of the information in interlinear glosses and translations that language documentation data for such treebanks typically comes with. In this paper, we improve upon such a bootstrapping model by representing glosses using a combination of morphological feature vectors and pre-trained lemma embeddings. We also contribute a mapping from glosses to Universal Dependencies morphological features.

Cite

CITATION STYLE

APA

Eibers, R., Evang, K., & Kallmeyer, L. (2023). Improving Low-resource RRG Parsing with Structured Gloss Embeddings. In FieldMatters 2023 - 2nd Workshop on NLP Applications to Field Linguistics, Proceedings (pp. 40–45). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.fieldmatters-1.6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free