Abstract
Value function estimation, i.e., prediction, is an important task in reinforcement learning. The Boltzmann softmax operator is a natural value estimator and can provide several benefits. However, it does not satisfy the non-expansion property, and its direct use may fail to converge even in value iteration. In this paper, we propose to update the value function with dynamic Boltzmann softmax (DBS) operator, which has good convergence property in the setting of planning and learning. Experimental results on GridWorld show that the DBS operator enables better estimation of the value function, which rectifies the convergence issue of the softmax operator. Finally, we propose the DBS-DQN algorithm by applying the DBS operator, which outperforms DQN substantially in 40 out of 49 Atari games.
Cite
CITATION STYLE
Pan, L., Cai, Q., Meng, Q., Chen, W., & Huang, L. (2020). Reinforcement learning with dynamic boltzmann softmax updates. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 1992–1998). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2020/276
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.