Abstract
Chinese NLP applications that rely on large text often contain huge amounts of vocabulary which are sparse in corpus. We show that characters' written form, Glyphs, in ideographic languages could carry rich semantics. We present a multi-modal model, Glyph2Vec, to tackle Chinese out-of-vocabulary word embedding problem. Glyph2Vec extracts visual features from word glyphs to expand current word embedding space for out-of-vocabulary word embedding, without the need of accessing any corpus, which is useful for improving Chinese NLP systems, especially for low-resource scenarios. Experiments across different applications show the significant effectiveness of our model.
Cite
CITATION STYLE
Chen, H. Y., Yu, S. H., & Lin, S. D. (2020). Glyph2Vec: Learning Chinese out-of-vocabulary word embedding from glyphs. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 2865–2871). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.256
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.