The multi-armed bandit is a reinforcement learning model where a learning agent re- peatedly chooses an action (pull a bandit arm) and the environment responds with a stochastic outcome (reward) coming from an unknown distribution associated with the chosen arm. Bandits have a wide-range of application such as Web recommendation sys- tems. We address the cumulative reward maximization problem in a secure federated learning setting, where multiple data owners keep their data stored locally and collaborate under the coordination of a central orchestration server. We rely on cryptographic schemes and propose Samba, a generic framework for Secure federAted Multi-armed BAndits. Each data owner has data associated to a bandit arm and the bandit algorithm has to sequen- tially select which data owner is solicited at each time step. We instantiate Samba for five bandit algorithms. We show that Samba returns the same cumulative reward as the non- secure versions of bandit algorithms, while satisfying formally proven security properties. We also show that the overhead due to cryptographic primitives is linear in the size of the input, which is confirmed by our proof-of-concept implementation.
CITATION STYLE
Ciucanu, R., Lafourcade, P., Marcadet, G., & Soare, M. (2022). SAMBA: A Generic Framework for Secure Federated Multi-Armed Bandits. Journal of Artificial Intelligence Research, 73, 737–765. https://doi.org/10.1613/JAIR.1.13163
Mendeley helps you to discover research relevant for your work.