Exclusive Feature Constrained Class Activation Mapping for Better Visual Explanation

12Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Whereas Deep Neural Network(DNN) shows wonderful performance on large scale data, lacking interpretability limits their usage in scenarios relevant to security. To make visual explanations less noisy and more class-discriminative, in this work, we propose a visual explanation method of DNN, named Exclusive Feature Constrained Class Activation Mapping(EFC-CAM). A new exclusive feature constraint is introduced to optimize the weight calculated from Grad-CAM or initialized from a constant vector. To better measure visual explanation methods, we design an effective evaluation metric which does not need bounding boxes as auxiliary information. Extensive quantitative experiments and visual inspection on ImageNet and Fashion validation set show the effectiveness of the proposed method.

Cite

CITATION STYLE

APA

Wang, P., Kong, X., Guo, W., & Zhang, X. (2021). Exclusive Feature Constrained Class Activation Mapping for Better Visual Explanation. IEEE Access, 9, 61417–61428. https://doi.org/10.1109/ACCESS.2021.3073465

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free