Explainable AI in drug discovery: self-interpretable graph neural network for molecular property prediction using concept whitening

28Citations
Citations of this article
47Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Molecular property prediction is a fundamental task in the field of drug discovery. Several works use graph neural networks to leverage molecular graph representations. Although they have been successfully applied in a variety of applications, their decision process is not transparent. In this work, we adapt concept whitening to graph neural networks. This approach is an explainability method used to build an inherently interpretable model, which allows identifying the concepts and consequently the structural parts of the molecules that are relevant for the output predictions. We test popular models on several benchmark datasets from MoleculeNet. Starting from previous work, we identify the most significant molecular properties to be used as concepts to perform classification. We show that the addition of concept whitening layers brings an improvement in both classification performance and interpretability. Finally, we provide several structural and conceptual explanations for the predictions.

Cite

CITATION STYLE

APA

Proietti, M., Ragno, A., Rosa, B. L., Ragno, R., & Capobianco, R. (2024). Explainable AI in drug discovery: self-interpretable graph neural network for molecular property prediction using concept whitening. Machine Learning, 113(4), 2013–2044. https://doi.org/10.1007/s10994-023-06369-y

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free