Distributed Reinforcement Learning for Multi-robot Decentralized Collective Construction

58Citations
Citations of this article
128Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Inspired by recent advances in single agent reinforcement learning, this paper extends the single-agent asynchronous advantage actor-critic (A3C) algorithm to enable multiple agents to learn a homogeneous, distributed policy, where agents work together toward a common goal without explicitly interacting. Our approach relies on centralized policy and critic learning, but decentralized policy execution, in a fully-observable system. We show that the sum of experience of all agents can be leveraged to quickly train a collaborative policy that naturally scales to smaller and larger swarms. We demonstrate the applicability of our method on a multi-robot construction problem, where agents need to arrange simple block elements to build a user-specified structure. We present simulation results where swarms of various sizes successfully construct different test structures without the need for additional training.

Cite

CITATION STYLE

APA

Sartoretti, G., Wu, Y., Paivine, W., Kumar, T. K. S., Koenig, S., & Choset, H. (2019). Distributed Reinforcement Learning for Multi-robot Decentralized Collective Construction. In Springer Proceedings in Advanced Robotics (Vol. 9, pp. 35–49). Springer Science and Business Media B.V. https://doi.org/10.1007/978-3-030-05816-6_3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free