Preemptive Image Robustification for Protecting Users against Man-in-the-Middle Adversarial Attacks

3Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

Deep neural networks have become the driving force of modern image recognition systems. However, the vulnerability of neural networks against adversarial attacks poses a serious threat to the people affected by these systems. In this paper, we focus on a real-world threat model where a Man-in-the-Middle adversary maliciously intercepts and perturbs images web users upload online. This type of attack can raise severe ethical concerns on top of simple performance degradation. To prevent this attack, we devise a novel bi-level optimization algorithm that finds points in the vicinity of natural images that are robust to adversarial perturbations. Experiments on CIFAR-10 and ImageNet show our method can effectively robustify natural images within the given modification budget. We also show the proposed method can improve robustness when jointly used with randomized smoothing.

Cite

CITATION STYLE

APA

Moon, S., An, G., & Song, H. O. (2022). Preemptive Image Robustification for Protecting Users against Man-in-the-Middle Adversarial Attacks. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 7823–7830). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i7.20751

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free