Flocking control of UAV swarms with deep reinforcement leaming approach

25Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The flocking control of UAV swarms has been studied extensively due to its wide applications. In this paper, the UAV flocking control problem is formulated as a Partial Observable Markov Decision Process (POMDP) where the constraints of the UAV's communication and perception ranges are considered. A deep reinforcement learning approach is proposed to solve this problem with centralized training and decentralized execution manner. The experience collected by all UAVs is used to train the shared flocking control policy, and each UAV performs actions based on the local environment information it observes. To enable the UAV swarm to maintain a flock and navigate in an environment with dense obstacles, a reward function is constructed considering with goal reaching, obstacles avoidance and flocking maintenance. Especially, the flocking maintenance reward is designed with the global information of the UAV swarm, which can only be obtained during the training phase. Simulation results demonstrate that the policy trained with the flocking maintenance reward can make the UAV swarm keep a flock when encountering obstacles and has good generalization ability with different number of UAVs.

Cite

CITATION STYLE

APA

Yan, P., Bai, C., Zheng, H., & Guo, J. (2020). Flocking control of UAV swarms with deep reinforcement leaming approach. In Proceedings of 2020 3rd International Conference on Unmanned Systems, ICUS 2020 (pp. 592–599). Institute of Electrical and Electronics Engineers Inc. https://doi.org/10.1109/ICUS50048.2020.9274899

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free