A study of qualitative knowledge-based exploration for continuous deep reinforcement learning

2Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

As an important method to solve sequential decisionmaking problems, reinforcement learning learns the policy of tasks through the interaction with environment. But it has difficulties scaling to largescale problems. One of the reasons is the exploration and exploitation dilemma which may lead to inefficient learning. We present an approach that addresses this shortcoming by introducing qualitative knowledge into reinforcement learning using cloud control systems to represent 'if-then' rules. We use it as the heuristics exploration strategy to guide the action selection in deep reinforcement learning. Empirical evaluation results show that our approach can make significant improvement in the learning process.

Cite

CITATION STYLE

APA

Li, C., Cao, L., Liu, X., Chen, X., Xu, Z., & Zhang, Y. (2017). A study of qualitative knowledge-based exploration for continuous deep reinforcement learning. IEICE Transactions on Information and Systems, E100D(11), 2721–2724. https://doi.org/10.1587/transinf.2017EDL8112

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free