Model-based reinforcement learning in continuous environments using real-time constrained optimization

13Citations
Citations of this article
94Readers
Mendeley users who have this article in their library.

Abstract

Reinforcement learning for robot control tasks in continuous environments is a challenging problem due to the dimensionality of the state and action spaces, time and resource costs for learning with a real robot as well as constraints imposed for its safe operation. In this paper we propose a modelbased reinforcement learning approach for continuous environments with constraints. The approach combines modelbased reinforcement learning with recent advances in approximate optimal control. This results in a bounded-rationality agent that makes decisions in real-time by efficiently solving a sequence of constrained optimization problems on learned sparse Gaussian process models. Such a combination has several advantages. No high-dimensional policy needs to be computed or stored while the learning problem often reduces to a set of lower-dimensional models of the dynamics. In addition, hard constraints can easily be included and objectives can also be changed in real-time to allow for multiple or dynamic tasks. The efficacy of the approach is demonstrated on both an extended cart pole domain and a challenging quadcopter navigation task using real data.

Cite

CITATION STYLE

APA

Andersson, O., Heintz, F., & Doherty, P. (2015). Model-based reinforcement learning in continuous environments using real-time constrained optimization. In Proceedings of the National Conference on Artificial Intelligence (Vol. 4, pp. 2497–2503). AI Access Foundation. https://doi.org/10.1609/aaai.v29i1.9623

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free