For task-nonspecific robot work in changing environment, most researches on developmental learning based on cognitive psychology advocate a staged developmental process and an explicit hierarchical action model. Though progress has been made by existing developmental learning approaches, there are still two open-ended inevitable problems: a) when numerous tasks are involved, the learning speed is not always satisfactory; b) when these tasks are not specified in advance, the hierarchical action model is hard to design beforehand or learn automatically. In order to solve these two problems, this paper proposes a new developmental reinforcement learning approach presented with its model and algorithms. In our model, any one of actor-critic learning models is encapsulated as a learning infrastructure to build an implicit action model called reward-policy mapping, and a self-motivated module is used for autonomous robots. The proposed approach efficaciously supports the implementation of an autonomous, interactive, cumulative and online learning process of task-nonspecific robots. The simulation results show that, to learn to perform nearly twenty thousand tasks, the proposed approach just needs half of the time that its counterpart, the actor-critic learning algorithm encapsulated, needs.
CITATION STYLE
Li, X., Yang, Y., Sun, Y., & Zhang, L. (2017). A developmental actor-critic reinforcement learning approach for task-nonspecific robot. In CGNCC 2016 - 2016 IEEE Chinese Guidance, Navigation and Control Conference (pp. 2231–2237). Institute of Electrical and Electronics Engineers Inc. https://doi.org/10.1109/CGNCC.2016.7829139
Mendeley helps you to discover research relevant for your work.