ENNclave: Offline Inference with Model Confidentiality

21Citations
Citations of this article
23Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Outsourcing machine learning inference creates a confidentiality dilemma: Either the client has to trust the server with potentially sensitive input data, or the server has to share his commercially valuable model. Known remedies include homomorphic encryption, multi-party computation, or placing the entire model in a trusted enclave. None of these are suitable for large models. For two relevant use cases, we show that it is possible to keep all confidential model parameters in the last (dense) layers of deep neural networks. This allows us to split the model such that the confidential parts fit into a trusted enclave on the client side. We present the eNNclave toolchain to cut TensorFlow models at any layer, splitting them into public and enclaved layers. This preserves TensorFlow's performance optimizations and hardware support for public layers, while keeping the parameters of the enclaved layers private. Evaluations on several machine learning tasks spanning multiple domains show that fast inference is possible while keeping the sensitive model parameters confidential. Accuracy results are close to the baseline where all layers carry sensitive information and confirm our approach is practical.

Cite

CITATION STYLE

APA

Schlögl, A., & Böhme, R. (2020). ENNclave: Offline Inference with Model Confidentiality. In AISec 2020 - Proceedings of the 13th ACM Workshop on Artificial Intelligence and Security (pp. 93–104). Association for Computing Machinery, Inc. https://doi.org/10.1145/3411508.3421376

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free