Deriving subgoals autonomously to accelerate learning in sparse reward domains

13Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.

Abstract

Sparse reward games, such as the infamous Montezuma's Revenge, pose a significant challenge for Reinforcement Learning (RL) agents. Hierarchical RL, which promotes efficient exploration via subgoals, has shown promise in these games. However, existing agents rely either on human domain knowledge or slow autonomous methods to derive suitable subgoals. In this work, we describe a new, autonomous approach for deriving subgoals from raw pixels that is more efficient than competing methods. We propose a novel intrinsic reward scheme for exploiting the derived subgoals, applying it to three Atari games with sparse rewards. Our agent's performance is comparable to that of state-of-the-art methods, demonstrating the usefulness of the subgoals found.

Cite

CITATION STYLE

APA

Dann, M., Zambetta, F., & Thangarajah, J. (2019). Deriving subgoals autonomously to accelerate learning in sparse reward domains. In 33rd AAAI Conference on Artificial Intelligence, AAAI 2019, 31st Innovative Applications of Artificial Intelligence Conference, IAAI 2019 and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019 (pp. 881–889). AAAI Press. https://doi.org/10.1609/aaai.v33i01.3301881

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free