Learning distributed word representations for bidirectional LSTM recurrent neural network

28Citations
Citations of this article
107Readers
Mendeley users who have this article in their library.

Abstract

Bidirectional long short-term memory (BLSTM) recurrent neural network (RNN) has been successfully applied in many tagging tasks. BLSTM-RNN relies on the distributed representation of words, which implies that the former can be futhermore improved through learning the latter better. In this work, we propose a novel approach to learn distributed word representations by training BLSTM-RNN on a specially designed task which only relies on unlabeled data. Our experimental results show that the proposed approach learns useful distributed word representations, as the trained representations significantly elevate the performance of BLSTM-RNN on three tagging tasks: part-of-speech tagging, chunking and named entity recognition, surpassing word representations trained by other published methods.

Cite

CITATION STYLE

APA

Wang, P., Qian, Y., Zhao, H., Soong, F. K., He, L., & Wu, K. (2016). Learning distributed word representations for bidirectional LSTM recurrent neural network. In 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL HLT 2016 - Proceedings of the Conference (pp. 527–533). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/n16-1064

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free