Explainable AI for Inspecting Adversarial Attacks on Deep Neural Networks

4Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Deep Neural Networks (DNN) are state of the art algorithms for image classification. Although significant achievements and perspectives, deep neural networks and accompanying learning algorithms have some important challenges to tackle. However, it appears that it is relatively easy to attack and fool with well-designed input samples called adversarial examples. Adversarial perturbations are unnoticeable for humans. Such attacks are a severe threat to the development of these systems in critical applications, such as medical or military systems. Hence, it is necessary to develop methods of counteracting these attacks. These methods are called defense strategies and aim at increasing the neural model’s robustness against adversarial attacks. In this paper, we reviewed the recent findings in adversarial attacks and defense strategies. We also analyzed the effects of attacks and defense strategies applied, using the local and global analyzing methods from the family of explainable artificial intelligence.

Cite

CITATION STYLE

APA

Klawikowska, Z., Mikołajczyk, A., & Grochowski, M. (2020). Explainable AI for Inspecting Adversarial Attacks on Deep Neural Networks. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12415 LNAI, pp. 134–146). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-61401-0_14

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free