Abstract
We introduce a novel mixed character-word architecture to improve Chinese sentence representations, by utilizing rich semantic information of word internal structures. Our architecture uses two key strategies. The first is a mask gate on characters, learning the relation among characters in a word. The second is a max-pooling operation on words, adaptively finding the optimal mixture of the atomic and compositional word representations. Finally, the proposed architecture is applied to various sentence composition models, which achieves substantial performance gains over baseline models on sentence similarity task.
Cite
CITATION STYLE
Wang, S., Zhang, J., & Zong, C. (2017). Exploiting word internal structures for generic Chinese sentence representation. In EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 298–303). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d17-1029
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.