Neural entity summarization with joint encoding and weak supervision

10Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

In a large-scale knowledge graph (KG), an entity is often described by a large number of triple-structured facts. Many applications require abridged versions of entity descriptions, called entity summaries. Existing solutions to entity summarization are mainly unsupervised. In this paper, we present a supervised approach NEST that is based on our novel neural model to jointly encode graph structure and text in KGs and generate high-quality diversified summaries. Since it is costly to obtain manually labeled summaries for training, our supervision is weak as we train with programmatically labeled data which may contain noise but is free of manual work. Evaluation results show that our approach significantly outperforms the state of the art on two public benchmarks.

Cite

CITATION STYLE

APA

Li, J., Cheng, G., Liu, Q., Zhang, W., Kharlamov, E., Gunaratna, K., & Chen, H. (2020). Neural entity summarization with joint encoding and weak supervision. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 1644–1650). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2020/228

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free