Harden Deep Convolutional Classifiers via K-Means Reconstruction

5Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Adversarial examples are carefully perturbed input examples that aim to mislead the deep neural network models into producing unexpected outputs. In this paper, we employ a K-means clustering algorithm as a pre-processing method to defend against adversarial examples. specifically, we reconstruct adversarial examples according to their cluster assignments in pixel level to reduce the impact of the injected perturbation. Our approach does not rely on any neural network architectures and can also work with existing pre-processing defenses to provide better protection for modern classifiers. Comprehensive comparison and evaluation have been conducted to investigate our proposal, where the models protected by the proposed defense show substantial robustness to strong adversarial attacks. As a by-product of our exploration of ensemble defense, we identify that the order of defense methods has a crucial impact on the final performance. Additionally, the limitation of K-means reconstruction and the impact of the number of clusters have also been studied to provide an in-deep understanding of pre-processing defenses.

Cite

CITATION STYLE

APA

Wang, F., He, L., Liu, W., & Zheng, Y. (2020). Harden Deep Convolutional Classifiers via K-Means Reconstruction. IEEE Access, 8, 168210–168218. https://doi.org/10.1109/ACCESS.2020.3024197

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free