Model-based reinforcement learning for approximate optimal control with temporal logic specifications

15Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper we study the problem of synthesizing optimal control policies for uncertain continuous-time nonlinear systems from syntactically co-safe linear temporal logic (scLTL) formulas. We formulate this problem as a sequence of reach-avoid optimal control sub-problems. We show that the resulting hybrid optimal control policy guarantees the satisfaction of a given scLTL formula by constructing a barrier certificate. Since solving each optimal control problem may be computationally intractable, we take a learning-based approach to approximately solve this sequence of optimal control problems online without requiring full knowledge of the system dynamics. Using Lyapunov-based tools, we develop sufficient conditions under which our approximate solution maintains correctness. Finally, we demonstrate the efficacy of the developed method with a numerical example.

Cite

CITATION STYLE

APA

Cohen, M. H., & Belta, C. (2021). Model-based reinforcement learning for approximate optimal control with temporal logic specifications. In HSCC 2021 - Proceedings of the 24th International Conference on Hybrid Systems: Computation and Control (part of CPS-IoT Week). Association for Computing Machinery, Inc. https://doi.org/10.1145/3447928.3456639

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free