Explaining Link Prediction Systems based on Knowledge Graph Embeddings

32Citations
Citations of this article
45Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Link Prediction (LP) aims at tackling Knowledge Graph incompleteness by inferring new, missing facts from the already known ones. The rise of novel Machine Learning techniques has led researchers to develop LP models that represent Knowledge Graph elements as vectors in an embedding space. These models can outperform traditional approaches and they can be employed in multiple downstream tasks; nonetheless, they tend to be opaque, and are mostly regarded as black boxes. Their lack of interpretability limits our understanding of their inner mechanisms, and undermines the trust that users can place in them. In this paper, we propose the novel Kelpie explainability framework. Kelpie can be applied to any embedding-based LP models independently from their architecture, and it explains predictions by identifying the combinations of training facts that have enabled them. Kelpie can extract two complementary types of explanations, that we dub necessary and sufficient. We describe in detail both the structure and the implementation details of Kelpie, and thoroughly analyze its performance through extensive experiments. Our results show that Kelpie significantly outperforms baselines across almost all scenarios.

Cite

CITATION STYLE

APA

Rossi, A., Firmani, D., Merialdo, P., & Teofili, T. (2022). Explaining Link Prediction Systems based on Knowledge Graph Embeddings. In Proceedings of the ACM SIGMOD International Conference on Management of Data (pp. 2062–2075). Association for Computing Machinery. https://doi.org/10.1145/3514221.3517887

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free