In this paper, we describe a method for predicting the understandability level of inferences with OWL. Specifically, we present a probabilistic model for measuring the understandability of a multiple-step inference based on the measurement of the understandability of individual inference steps. We also present an evaluation study which confirms that our model works relatively well for two-step inferences with OWL. This model has been applied in our research on generating accessible explanations for an entailment of OWL ontologies, to determine the most understandable inference among alternatives, from which the final explanation is generated. © 2013 Springer-Verlag Berlin Heidelberg.
CITATION STYLE
Nguyen, T. A. T., Power, R., Piwek, P., & Williams, S. (2013). Predicting the understandability of OWL inferences. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7882 LNCS, pp. 109–123). Springer Verlag. https://doi.org/10.1007/978-3-642-38288-8_8
Mendeley helps you to discover research relevant for your work.