Decomposing user-defined tasks in a reinforcement learning setup using TextWorld

0Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.

Abstract

The current paper proposes a hierarchical reinforcement learning (HRL) method to decompose a complex task into simpler sub-tasks and leverage those to improve the training of an autonomous agent in a simulated environment. For practical reasons (i.e., illustrating purposes, easy implementation, user-friendly interface, and useful functionalities), we employ two Python frameworks called TextWorld and MiniGrid. MiniGrid functions as a 2D simulated representation of the real environment, while TextWorld functions as a high-level abstraction of this simulated environment. Training on this abstraction disentangles manipulation from navigation actions and allows us to design a dense reward function instead of a sparse reward function for the lower-level environment, which, as we show, improves the performance of training. Formal methods are utilized throughout the paper to establish that our algorithm is not prevented from deriving solutions.

Cite

CITATION STYLE

APA

Petsanis, T., Keroglou, C., Ch. Kapoutsis, A., Kosmatopoulos, E. B., & Sirakoulis, G. C. (2023). Decomposing user-defined tasks in a reinforcement learning setup using TextWorld. Frontiers in Robotics and AI, 10. https://doi.org/10.3389/frobt.2023.1280578

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free