Search space of adversarial perturbations against image filters

0Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

Abstract

The superiority of deep learning performance is threatened by safety issues for itself. Recent findings have shown that deep learning systems are very weak to adversarial examples, an attack form that was altered by the attacker's intent to deceive the deep learning system. There are many proposed defensive methods to protect deep learning systems against adversarial examples. However, there is still lack of principal strategies to deceive those defensive methods. Any time a particular countermeasure is proposed, a new powerful adversarial attack will be invented to deceive that countermeasure. In this study, we focus on investigating the ability to create adversarial patterns in search space against defensive methods that use image filters. Experimental results conducted on the ImageNet dataset with image classification tasks showed the correlation between the search space of adversarial perturbation and filters. These findings open a new direction for building stronger offensive methods towards deep learning systems.

Cite

CITATION STYLE

APA

Thang, D. D., & Matsui, T. (2020). Search space of adversarial perturbations against image filters. International Journal of Advanced Computer Science and Applications, 11(1), 11–19. https://doi.org/10.14569/ijacsa.2020.0110102

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free