Generalizing Reinforcement Learning through Fusing Self-Supervised Learning into Intrinsic Motivation

6Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

Despite the great potential of reinforcement learning (RL) in solving complex decision-making problems, generalization remains one of its key challenges, leading to difficulty in deploying learned RL policies to new environments. In this paper, we propose to improve the generalization of RL algorithms through fusing Self-supervised learning into Intrinsic Motivation (SIM). Specifically, SIM boosts representation learning through driving the cross-correlation matrix between the embeddings of augmented and non-augmented samples close to the identity matrix. This aims to increase the similarity between the embedding vectors of a sample and its augmented version while minimizing the redundancy between the components of these vectors. Meanwhile, the redundancy reduction based self-supervised loss is converted to an intrinsic reward to further improve generalization in RL via an auxiliary objective. As a general paradigm, SIM can be implemented on top of any RL algorithm. Extensive evaluations have been performed on a diversity of tasks. Experimental results demonstrate that SIM consistently outperforms the state-of-the-art methods and exhibits superior generalization capability and sample efficiency.

Cite

CITATION STYLE

APA

Wu, K., Wu, M., Chen, Z., Xu, Y., & Li, X. (2022). Generalizing Reinforcement Learning through Fusing Self-Supervised Learning into Intrinsic Motivation. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 8683–8690). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i8.20847

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free