Diffchaser: Detecting disagreements for deep neural networks

63Citations
Citations of this article
56Readers
Mendeley users who have this article in their library.

Abstract

The platform migration and customization have become an indispensable process of deep neural network (DNN) development lifecycle. A high-precision but complex DNN trained in the cloud on massive data and powerful GPUS often goes through an optimization phase (e.g., quantization, compression) before deployment to a target device (e.g., mobile device). A test set that effectively uncovers the disagreements of a DNN and its optimized variant provides certain feedback to debug and further enhance the optimization procedure. However, the minor inconsistency between a DNN and its optimized version is often hard to detect and easily bypasses the original test set. This paper proposes DiffChaser, an automated black-box testing framework to detect untargeted/targeted disagreements between version variants of a DNN. We demonstrate 1) its effectiveness by comparing with the state-of-the-art techniques, and 2) its usefulness in real-world DNN product deployment involved with quantization and optimization.

Cite

CITATION STYLE

APA

Xie, X., Ma, L., Wang, H., Li, Y., Liu, Y., & Li, X. (2019). Diffchaser: Detecting disagreements for deep neural networks. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2019-August, pp. 5772–5778). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2019/800

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free