A simple and effective method for injecting word-level information into character-aware neural language models

2Citations
Citations of this article
60Readers
Mendeley users who have this article in their library.

Abstract

We propose a simple and effective method to inject word-level information into character-aware neural language models. Unlike previous approaches which usually inject wordlevel information at the input of a long short-term memory (LSTM) network, we inject it into the softmax function. The resultant model can be seen as a combination of character-aware language model and simple word-level language model. Our injection method can also be used together with previous methods. Through the experiments on 14 typologically diverse languages, we empirically show that our injection method, when used together with the previous methods, works better than the previous methods, including a gating mechanism, averaging, and concatenation of word vectors. We also provide a comprehensive comparison of these injection methods.

Cite

CITATION STYLE

APA

Feng, Y., Kamigaito, H., Takamura, H., & Okumura, M. (2019). A simple and effective method for injecting word-level information into character-aware neural language models. In CoNLL 2019 - 23rd Conference on Computational Natural Language Learning, Proceedings of the Conference (pp. 920–928). Association for Computational Linguistics. https://doi.org/10.5715/jnlp.30.156

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free