Deep Reinforcement Learning with Hidden Layers on Future States

0Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Deep reinforcement learning algorithms such as Deep Q-Networks have successfully been used to construct a strong agent for Atari games by only performing direct evaluation of the current state and actions. This is in stark contrast to the algorithms for traditional board games such as Chess or Go, where a look-ahead search mechanism is indispensable to build a strong agent. In this paper, we present a novel deep reinforcement learning architecture that can both effectively and efficiently use information on future states in video games. First, we demonstrate that such information is indeed quite useful in deep reinforcement learning by using exact state transition information obtained from the emulator. We then propose a method that predicts future states using Long Short Term Memory (LSTM), such that the agent can look ahead without the emulator. In this work, we applied our method to the asynchronous advantage actor-critic (A3C) architecture. The experimental results show that our proposed method with predicted future states substantially outperforms the vanilla A3C in several Atari games.

Cite

CITATION STYLE

APA

Kameko, H., Suzuki, J., Mizukami, N., & Tsuruoka, Y. (2018). Deep Reinforcement Learning with Hidden Layers on Future States. In Communications in Computer and Information Science (Vol. 818, pp. 46–60). Springer Verlag. https://doi.org/10.1007/978-3-319-75931-9_4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free