Neuro-Symbolic Approaches for Text-Based Policy Learning

8Citations
Citations of this article
61Readers
Mendeley users who have this article in their library.

Abstract

Text-Based Games (TBGs) have emerged as important testbeds for reinforcement learning (RL) in the natural language domain. Previous methods using LSTM-based action policies are uninterpretable and often overfit the training games showing poor performance to unseen test games. We present SymboLic Action policy for Textual Environments (SLATE), that learns interpretable action policy rules from symbolic abstractions of textual observations for improved generalization. We outline a method for end-to-end differentiable symbolic rule learning and show that such symbolic policies outperform previous state-of-the-art methods in text-based RL for the coin collector environment from 5−10x fewer training games. Additionally, our method provides human-understandable policy rules that can be readily verified for their logical consistency and can be easily debugged.

Cite

CITATION STYLE

APA

Chaudhury, S., Sen, P., Ono, M., Kimura, D., Tatsubori, M., & Munawar, A. (2021). Neuro-Symbolic Approaches for Text-Based Policy Learning. In EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 3073–3078). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.emnlp-main.245

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free