Differentially Private Stochastic Coordinate Descent

7Citations
Citations of this article
25Readers
Mendeley users who have this article in their library.

Abstract

In this paper we tackle the challenge of making the stochastic coordinate descent algorithm differentially private. Compared to the classical gradient descent algorithm where updates operate on a single model vector and controlled noise addition to this vector suffices to hide critical information about individuals, stochastic coordinate descent crucially relies on keeping auxiliary information in memory during training. This auxiliary information provides an additional privacy leak and poses the major challenge addressed in this work. Driven by the insight that under independent noise addition, the consistency of the auxiliary information holds in expectation, we present DP-SCD, the first differentially private stochastic coordinate descent algorithm. We analyze our new method theoretically and argue that decoupling and parallelizing coordinate updates is essential for its utility. On the empirical side we demonstrate competitive performance against the popular stochastic gradient descent alternative (DP-SGD) while requiring significantly less tuning.

Cite

CITATION STYLE

APA

Damaskinos, G., Mendler-Dünner, C., Guerraoui, R., Papandreou, N., & Parnell, T. (2021). Differentially Private Stochastic Coordinate Descent. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 8B, pp. 7176–7184). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i8.16882

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free