Learning to act optimally in partially observable Markov decision processes using hybrid probabilistic logic programs

0Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a probabilistic logic programming framework to reinforcement learning, by integrating reinforcement learning, in POMDP environments, with normal hybrid probabilistic logic programs with probabilistic answer set semantics, that is capable of representing domain-specific knowledge. We formally prove the correctness of our approach. We show that the complexity of finding a policy for a reinforcement learning problem in our approach is NP-complete. In addition, we show that any reinforcement learning problem can be encoded as a classical logic program with answer set semantics. We also show that a reinforcement learning problem can be encoded as a SAT problem. We present a new high level action description language that allows the factored representation of POMDP. Moreover, we modify the original model of POMDP so that it be able to distinguish between knowledge producing actions and actions that change the environment. © 2011 Springer-Verlag.

Cite

CITATION STYLE

APA

Saad, E. (2011). Learning to act optimally in partially observable Markov decision processes using hybrid probabilistic logic programs. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6929 LNAI, pp. 504–519). https://doi.org/10.1007/978-3-642-23963-2_39

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free