Scalable Partial Explainability in Neural Networks via Flexible Activation Functions (Student Abstract)

6Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

Abstract

Current state-of-the-art neural network explanation methods (e.g. Saliency maps, DeepLIFT, LIME, etc.) focus more on the direct relationship between NN outputs and inputs rather than the NN structure and operations itself, hence there still exists uncertainty over the exact role played by neurons. In this paper, we propose a novel neural network structure with Kolmogorov-Arnold Superposition Theorem based topology and Gaussian Processes based flexible activation function to achieve partial explainability of the neuron inner reasoning. The model feasibility is verified in a case study on binary classification of the banknotes.

Cite

CITATION STYLE

APA

Sun, S. C., Li, C., Wei, Z., Tsourdos, A., & Guo, W. (2021). Scalable Partial Explainability in Neural Networks via Flexible Activation Functions (Student Abstract). In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 18, pp. 15899–15900). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i18.17946

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free