SAMBA: A Generic Framework for Secure Federated Multi-Armed Bandits

6Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

The multi-armed bandit is a reinforcement learning model where a learning agent re- peatedly chooses an action (pull a bandit arm) and the environment responds with a stochastic outcome (reward) coming from an unknown distribution associated with the chosen arm. Bandits have a wide-range of application such as Web recommendation sys- tems. We address the cumulative reward maximization problem in a secure federated learning setting, where multiple data owners keep their data stored locally and collaborate under the coordination of a central orchestration server. We rely on cryptographic schemes and propose Samba, a generic framework for Secure federAted Multi-armed BAndits. Each data owner has data associated to a bandit arm and the bandit algorithm has to sequen- tially select which data owner is solicited at each time step. We instantiate Samba for five bandit algorithms. We show that Samba returns the same cumulative reward as the non- secure versions of bandit algorithms, while satisfying formally proven security properties. We also show that the overhead due to cryptographic primitives is linear in the size of the input, which is confirmed by our proof-of-concept implementation.

Cite

CITATION STYLE

APA

Ciucanu, R., Lafourcade, P., Marcadet, G., & Soare, M. (2022). SAMBA: A Generic Framework for Secure Federated Multi-Armed Bandits. Journal of Artificial Intelligence Research, 73, 737–765. https://doi.org/10.1613/JAIR.1.13163

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free