Abstract
Character-aware neural language models can capture the relationship between words by exploiting character-level information and are particularly effective for languages with rich morphology. However, these models are usually biased towards information from surface forms. To alleviate this problem, we propose a simple and effective method to improve a character-aware neural language model by forcing a character encoder to produce word-based embeddings under Skip-gram architecture in a warm-up step without extra training data. We empirically show that the resulting character-aware neural language model achieves obvious improvements of perplexity scores on typologically diverse languages, that contain many low-frequency or unseen words.
Cite
CITATION STYLE
Feng, Y., Hu, C., Kamigaito, H., Takamura, H., & Okumura, M. (2021). Improving Character-Aware Neural Language Model by Warming Up Character Encoder under Skip-gram Architecture. In International Conference Recent Advances in Natural Language Processing, RANLP (pp. 421–427). Incoma Ltd. https://doi.org/10.26615/978-954-452-072-4_048
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.