Multi-agent Pathfinding with Communication Reinforcement Learning and Deadlock Detection

5Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The learning-based approach has been proved to be an effective way to solve multi-agent path finding (MAPF) problems. For large warehouse systems, the distributed strategy based on learning method can effectively improve efficiency and scalability. But compared with the traditional centralized planner, the learning-based approach is more prone to deadlocks. Communication learning has also made great progress in the field of multi-agent in recent years and has been be introduced into MAPF. However, the current communication methods provide redundant information for reinforcement learning and interfere with the decision-making of agents. In this paper, we combine the reinforcement learning with communication learning. The agents select its communication objectives based on priority and mask off redundant communication links. Then we use a feature interactive network based on graph neural network to achieve the information aggregation. We also introduce an additional deadlock detection mechanism to increase the likelihood of an agent escaping a deadlock. Experiments demonstrate our method is able to plan collision-free paths in different warehouse environments.

Cite

CITATION STYLE

APA

Ye, Z., Li, Y., Guo, R., Gao, J., & Fu, W. (2022). Multi-agent Pathfinding with Communication Reinforcement Learning and Deadlock Detection. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13455 LNAI, pp. 493–504). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-13844-7_47

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free