I4R: Promoting deep reinforcement learning by the indicator for expressive representations

7Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

Abstract

Learning expressive representations is always crucial for well-performed policies in deep reinforcement learning (DRL). Different from supervised learning, in DRL, accurate targets are not always available, and some inputs with different actions only have tiny differences, which stimulates the demand for learning expressive representations. In this paper, firstly, we empirically compare the representations of DRL models with different performances. We observe that the representations of a better state extractor (SE) are more scattered than a worse one when they are visualized. Thus, we investigate the singular values of representation matrix, and find that, better SEs always correspond to smaller differences among these singular values. Next, based on such observations, we define an indicator of the representations for DRL model, which is the Number of Significant Singular Values (NSSV) of a representation matrix. Then, we propose I4R algorithm, to improve DRL algorithms by adding the corresponding regularization term to enhance the NSSV. Finally, we apply I4R to both policy gradient and value based algorithms on Atari games, and the results show the superiority of our proposed method.1,.

Cite

CITATION STYLE

APA

Luo, X., Meng, Q., He, D., Chen, W., & Wang, Y. (2020). I4R: Promoting deep reinforcement learning by the indicator for expressive representations. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 2669–2675). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2020/370

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free