Towards certificated model robustness against weight perturbations

24Citations
Citations of this article
22Readers
Mendeley users who have this article in their library.

Abstract

This work studies the sensitivity of neural networks to weight perturbations, firstly corresponding to a newly developed threat model that perturbs the neural network parameters. We propose an efficient approach to compute a certified robustness bound of weight perturbations, within which neural networks will not make erroneous outputs as desired by the adversary. In addition, we identify a useful connection between our developed certification method and the problem of weight quantization, a popular model compression technique in deep neural networks (DNNs) and a 'must-try' step in the design of DNN inference engines on resource constrained computing platforms, such as mobiles, FPGA, and ASIC. Specifically, we study the problem of weight quantization - weight perturbations in the non-adversarial setting - through the lens of certificated robustness, and we demonstrate significant improvements on the generalization ability of quantized networks through our robustness-aware quantization scheme.

Cite

CITATION STYLE

APA

Weng, T. W., Zhao, P., Liu, S., Chen, P. Y., Lin, X., & Daniel, L. (2020). Towards certificated model robustness against weight perturbations. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 6356–6363). AAAI press. https://doi.org/10.1609/aaai.v34i04.6105

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free