Convolutional Neural Network Visualization in Adversarial Example Attack

0Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In deep learning, repeated convolution and pooling processes help to learn image features, but complex nonlinear operations make deep learning models difficult for users to understand. Adversarial example attack is a unique form of attack in deep learning. The attacker attacks the model by applying invisible changes to the picture, affecting the results of the model judgment. In this paper, a research is implemented on the adversarial example attack and neural network interpretability. The neural network interpretability research is believed to have considerable potential in resisting adversarial examples. It helped understand how the adversarial examples induce the neural network to make a wrong judgment and identify adversarial examples in the test set. The corresponding algorithm was designed and the image recognition model was built based on the ImageNet training set. And then the adversarial-example generation algorithm and the neural network visualization algorithm were designed to determine the model learning heat map of the original example and the adversarial-example. The results show that it develops the application of neural network interpretability in the field of resisting adversarial-example attacks.

Cite

CITATION STYLE

APA

Yu, C., Wang, X., & Li, Y. (2020). Convolutional Neural Network Visualization in Adversarial Example Attack. In Communications in Computer and Information Science (Vol. 1257 CCIS, pp. 247–258). Springer. https://doi.org/10.1007/978-981-15-7981-3_16

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free