Collaborative multi-agent reinforcement learning based on experience propagation

7Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

For multi-agent reinforcement learning in Markov games, knowledge extraction and sharing are key research probl ems. State list extracting means to calculate the optimal shared state path from state trajectories with cycles. A state list extracting algorithm checks cyclic state lists of a current state in the state traj ectory, condensing the optimal action set of the current state. By reinforcing the optimal action selected, the action policy of cyclic states is optimized gradually. The state list extracting is repeatedly learned and used as the experience knowledge which is shared by teams. Agents speed up the rate of convergence by experience sharing. Competition games of preys and predators are used for the experiments. The results of experiments prove that the prop osed algorithms overcome the lack of experience in the initial stage, speed up learning and improve the performance.

Cite

CITATION STYLE

APA

Fang, M., & Groen, F. C. A. (2013). Collaborative multi-agent reinforcement learning based on experience propagation. Journal of Systems Engineering and Electronics, 24(4), 683–689. https://doi.org/10.1109/JSEE.2013.00079

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free