Faithful Embeddings for EL+ + Knowledge Bases

11Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recently, increasing efforts are put into learning continual representations for symbolic knowledge bases (KBs). However, these approaches either only embed the data-level knowledge (ABox) or suffer from inherent limitations when dealing with concept-level knowledge (TBox), i.e., they cannot faithfully model the logical structure present in the KBs. We present BoxEL, a geometric KB embedding approach that allows for better capturing the logical structure (i.e., ABox and TBox axioms) in the description logic EL+ +. BoxEL models concepts in a KB as axis-parallel boxes that are suitable for modeling concept intersection, entities as points inside boxes, and relations between concepts/entities as affine transformations. We show theoretical guarantees (soundness) of BoxEL for preserving logical structure. Namely, the learned model of BoxEL embedding with loss 0 is a (logical) model of the KB. Experimental results on (plausible) subsumption reasonings and a real-world application–protein-protein prediction show that BoxEL outperforms traditional knowledge graph embedding methods as well as state-of-the-art EL+ + embedding approaches.

Cite

CITATION STYLE

APA

Xiong, B., Potyka, N., Tran, T. K., Nayyeri, M., & Staab, S. (2022). Faithful Embeddings for EL+ + Knowledge Bases. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13489 LNCS, pp. 22–38). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-19433-7_2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free