Improving Character-Aware Neural Language Model by Warming Up Character Encoder under Skip-gram Architecture

3Citations
Citations of this article
40Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Character-aware neural language models can capture the relationship between words by exploiting character-level information and are particularly effective for languages with rich morphology. However, these models are usually biased towards information from surface forms. To alleviate this problem, we propose a simple and effective method to improve a character-aware neural language model by forcing a character encoder to produce word-based embeddings under Skip-gram architecture in a warm-up step without extra training data. We empirically show that the resulting character-aware neural language model achieves obvious improvements of perplexity scores on typologically diverse languages, that contain many low-frequency or unseen words.

Cite

CITATION STYLE

APA

Feng, Y., Hu, C., Kamigaito, H., Takamura, H., & Okumura, M. (2021). Improving Character-Aware Neural Language Model by Warming Up Character Encoder under Skip-gram Architecture. In International Conference Recent Advances in Natural Language Processing, RANLP (pp. 421–427). Incoma Ltd. https://doi.org/10.26615/978-954-452-072-4_048

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free