structured information representing knowledge encoded in scientific publications. With the sheer volume of published scientific literature comprising a plethora of inhomogeneous entities and relations to describe scientific concepts, these KGs are inherently incomplete. We present exBERT, a method for leveraging pre-trained transformer language models to perform scholarly knowledge graph completion. We model triples of a knowledge graph as text and perform triple classification (i.e., belongs to KG or not). The evaluation shows that exBERT outperforms other baselines on three scholarly KG completion datasets in the tasks of triple classification, link prediction, and relation prediction. Furthermore, we present two scholarly datasets as resources for the research community, collected from public KGs and online resources.
CITATION STYLE
Jaradeh, M. Y., Singh, K., Stocker, M., & Auer, S. (2021). Triple Classification for Scholarly Knowledge Graph Completion. In K-CAP 2021 - Proceedings of the 11th Knowledge Capture Conference (pp. 225–232). Association for Computing Machinery, Inc. https://doi.org/10.1145/3460210.3493582
Mendeley helps you to discover research relevant for your work.