Abstract
Model explanations are generated by XAI (explainable AI)methods to help people understand and interpret machine learning models. To study XAI methods from the human perspective, we propose a human-based benchmark dataset, i.e., human saliency benchmark (HSB), for evaluating saliency based XAI methods. Different from existing human saliency annotations where class-related features are manually and subjectively labeled, this benchmark collects more objective human attention on vision information with a precise eyetracking device and a novel crowdsourcing experiment. Taking the labor cost of human experiment into consideration, we further explore the potential of utilizing a prediction model trained on HSB to mimic saliency annotating by humans. Hence, a dense prediction problem is formulated, and we propose an encoder-decoder architecture which combines multimodal and multi-scale features to produce the human saliency maps. Accordingly, a pretraining-finetuning method is designed to address the model training problem. Finally, we arrive at a model trained on HSB named human saliency imitator(HSI).We show, through an extensive evaluation, that HIS can successfully predict human saliency on our HSB dataset, and the HSI-generated human saliency dataset on Image Netshow cases the ability of benchmarking XAI methods both qualitatively and quantitatively.
Cite
CITATION STYLE
Yang, Y., Zheng, Y., Deng, D., Zhang, J., Huang, Y., Yang, Y., … Cao, C. C. (2022). HSI: Human Saliency Imitator for Benchmarking Saliency-Based Model Explanations. In Proceedings of the AAAI Conference on Human Computation and Crowdsourcing (Vol. 10, pp. 231–242). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/hcomp.v10i1.22002
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.