Deep reinforcement learning provides a promising approach for text-based games in studying natural language communication between humans and artificial agents. However, the generalization still remains a big challenge as the agents depend critically on the complexity and variety of training tasks. In this paper, we address this problem by introducing a hierarchical framework built upon the knowledge graph-based RL agent. In the high level, a meta-policy is executed to decompose the whole game into a set of subtasks specified by textual goals, and select one of them based on the KG. Then a subpolicy in the low level is executed to conduct goal-conditioned reinforcement learning. We carry out experiments on games with various difficulty levels and show that the proposed method enjoys favorable generalizability.
CITATION STYLE
Xu, Y., Fang, M., Chen, L., Du, Y., & Zhang, C. (2021). Generalization in Text-based Games via Hierarchical Reinforcement Learning. In Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 (pp. 1343–1353). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-emnlp.116
Mendeley helps you to discover research relevant for your work.