Abstract
Treebanking for local languages is hampered by the lack of existing parsers to generate pre-annotations. However, it has been shown that reasonably accurate parsers can be bootstrapped with little initial training data when use is made of the information in interlinear glosses and translations that language documentation data for such treebanks typically comes with. In this paper, we improve upon such a bootstrapping model by representing glosses using a combination of morphological feature vectors and pre-trained lemma embeddings. We also contribute a mapping from glosses to Universal Dependencies morphological features.
Cite
CITATION STYLE
Eibers, R., Evang, K., & Kallmeyer, L. (2023). Improving Low-resource RRG Parsing with Structured Gloss Embeddings. In FieldMatters 2023 - 2nd Workshop on NLP Applications to Field Linguistics, Proceedings (pp. 40–45). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.fieldmatters-1.6
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.