Multi Agent Deep Learning with Cooperative Communication

8Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

We consider the problem of multi agents cooperating in a partially-observable environment. Agents must learn to coordinate and share relevant information to solve the tasks successfully. This article describes Asynchronous Advantage Actor-Critic with Communication (A3C2), an end-to-end differentiable approach where agents learn policies and communication protocols simultaneously. A3C2 uses a centralized learning, distributed execution paradigm, supports independent agents, dynamic team sizes, partially-observable environments, and noisy communications. We compare and show that A3C2 outperforms other state-of-the-art proposals in multiple environments.

Cite

CITATION STYLE

APA

Simões, D., Lau, N., & Reis, L. P. (2020). Multi Agent Deep Learning with Cooperative Communication. Journal of Artificial Intelligence and Soft Computing Research, 10(3), 189–207. https://doi.org/10.2478/jaiscr-2020-0013

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free