Optimal and efficient stochastic motion planning in partially-known environments

12Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

A framework capable of computing optimal control policies for a continuous system in the presence of both action and environment uncertainty is presented in this work. The framework decomposes the planning problem into two stages: an offline phase that reasons only over action uncertainty and an online phase that quickly reacts to the uncertain environment. Offline, a bounded-parameter Markov decision process (BMDP) is employed to model the evolution of the stochastic sys-tem over a discretization of the environment. Online, an optimal control policy over the BMDP is computed. Upon the discovery of an unknown environment feature during policy execution, the BMDP is updated and the optimal control policy is efficiently recomputed. Depending on the desired quality of the control policy, a suite of methods is presented to incorporate new information into the BMDP with varying degrees of detail online. Experiments confirm that the framework recomputes high-quality policies in seconds and is orders of magnitude faster than existing methods.

Cite

CITATION STYLE

APA

Luna, R., Lahijanian, M., Moll, M., & Kavraki, L. E. (2014). Optimal and efficient stochastic motion planning in partially-known environments. In Proceedings of the National Conference on Artificial Intelligence (Vol. 4, pp. 2549–2555). AI Access Foundation. https://doi.org/10.1609/aaai.v28i1.9054

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free