Secure aggregation for federated learning in flower

33Citations
Citations of this article
31Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Federated Learning (FL) allows parties to learn a shared prediction model by delegating the training computation to clients and aggregating all the separately trained models on the server. To prevent private information being inferred from local models, Secure Aggregation (SA) protocols are used to ensure that the server is unable to inspect individual trained models as it aggregates them. However, current implementations of SA in FL frameworks have limitations, including vulnerability to client dropouts or configuration difficulties. In this paper, we present Salvia, an implementation of SA for Python users in the Flower FL framework. Based on the SecAgg(+) protocols for a semi-honest threat model, Salvia is robust against client dropouts and exposes a flexible and easy-to-use API that is compatible with various machine learning frameworks. We show that Salvia's experimental performance is consistent with SecAgg(+)'s theoretical computation and communication complexities.

Cite

CITATION STYLE

APA

Li, K. H., De Gusmão, P. P. B., Beutel, D. J., & Lane, N. D. (2021). Secure aggregation for federated learning in flower. In DistributedML 2021 - Proceedings of the 2nd ACM International Workshop on Distributed Machine Learning, Part of CoNEXT 2021 (pp. 8–14). Association for Computing Machinery, Inc. https://doi.org/10.1145/3488659.3493776

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free