Reinforcement learning with dynamic boltzmann softmax updates

26Citations
Citations of this article
43Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Value function estimation, i.e., prediction, is an important task in reinforcement learning. The Boltzmann softmax operator is a natural value estimator and can provide several benefits. However, it does not satisfy the non-expansion property, and its direct use may fail to converge even in value iteration. In this paper, we propose to update the value function with dynamic Boltzmann softmax (DBS) operator, which has good convergence property in the setting of planning and learning. Experimental results on GridWorld show that the DBS operator enables better estimation of the value function, which rectifies the convergence issue of the softmax operator. Finally, we propose the DBS-DQN algorithm by applying the DBS operator, which outperforms DQN substantially in 40 out of 49 Atari games.

Cite

CITATION STYLE

APA

Pan, L., Cai, Q., Meng, Q., Chen, W., & Huang, L. (2020). Reinforcement learning with dynamic boltzmann softmax updates. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 1992–1998). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2020/276

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free