Model-based reinforcement learning for approximate optimal control with temporal logic specifications

From MaRDI portal
Publication:6201592

DOI10.1145/3447928.3456639arXiv2101.07156WikidataQ130844063 ScholiaQ130844063MaRDI QIDQ6201592FDOQ6201592


Authors: M. Cohen, Calin Belta Edit this on Wikidata


Publication date: 21 February 2024

Published in: Proceedings of the 24th International Conference on Hybrid Systems: Computation and Control (Search for Journal in Brave)

Abstract: In this paper we study the problem of synthesizing optimal control policies for uncertain continuous-time nonlinear systems from syntactically co-safe linear temporal logic (scLTL) formulas. We formulate this problem as a sequence of reach-avoid optimal control sub-problems. We show that the resulting hybrid optimal control policy guarantees the satisfaction of a given scLTL formula by constructing a barrier certificate. Since solving each optimal control problem may be computationally intractable, we take a learning-based approach to approximately solve this sequence of optimal control problems online without requiring full knowledge of the system dynamics. Using Lyapunov-based tools, we develop sufficient conditions under which our approximate solution maintains correctness. Finally, we demonstrate the efficacy of the developed method with a numerical example.


Full work available at URL: https://arxiv.org/abs/2101.07156







Cites Work


Cited In (1)





This page was built for publication: Model-based reinforcement learning for approximate optimal control with temporal logic specifications

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6201592)