Abstract
Knowledge bases are useful resources for many natural language processing tasks, however, they are far from complete. In this paper, we define a novel entity representation as a mixture of its neighborhood in the knowledge base and apply this technique on TransE—a well-known embedding model for knowledge base completion. Experimental results show that the neighborhood information significantly helps to improve the results of the TransE, leading to better performance than obtained by other state-of-the-art embedding models on three benchmark datasets for triple classification, entity prediction and relation prediction tasks.
Author supplied keywords
Cite
CITATION STYLE
Nguyen, D. Q., Sirts, K., Qu, L., & Johnson, M. (2016). Neighborhood mixture model for knowledge base completion. In CoNLL 2016 - 20th SIGNLL Conference on Computational Natural Language Learning, Proceedings (pp. 40–50). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/k16-1005
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.