PID-Based Approach to Adversarial Attacks

17Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

Abstract

Adversarial attack can misguide the deep neural networks (DNNs) with adding small-magnitude perturbations to normal examples, which is mainly determined by the gradient of the loss function with respect to inputs. Previously, various strategies have been proposed to enhance the performance of adversarial attacks. However, all these methods only utilize the gradients in the present and past to generate adversarial examples. Until now, the trend of gradient change in the future (i.e., the derivative of gradient) has not been considered yet. Inspired by the classic proportional-integral-derivative (PID) controller in the field of automatic control, we propose a new PID-based approach for generating adversarial examples. The gradients in the present and past, and the derivative of gradient are considered in our method, which correspond to the components of P, I and D in the PID controller, respectively. Extensive experiments consistently demonstrate that our method can achieve higher attack success rates and exhibit better transferability compared with the state-of-the-art gradient-based adversarial attacks. Furthermore, our method possesses good extensibility and can be applied to almost all available gradient-based adversarial attacks.

Cite

CITATION STYLE

APA

Wan, C., Ye, B., & Huang, F. (2021). PID-Based Approach to Adversarial Attacks. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 11B, pp. 10033–10040). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i11.17204

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free