Safe exploration in model-based reinforcement learning using control barrier functions

From MaRDI portal
Publication:2103658

DOI10.1016/J.AUTOMATICA.2022.110684zbMATH Open1505.93123arXiv2104.08171OpenAlexW3154507809MaRDI QIDQ2103658FDOQ2103658


Authors: M. Cohen, Calin Belta Edit this on Wikidata


Publication date: 9 December 2022

Published in: Automatica (Search for Journal in Brave)

Abstract: This paper develops a model-based reinforcement learning (MBRL) framework for learning online the value function of an infinite-horizon optimal control problem while obeying safety constraints expressed as control barrier functions (CBFs). Our approach is facilitated by the development of a novel class of CBFs, termed Lyapunov-like CBFs (LCBFs), that retain the beneficial properties of CBFs for developing minimally-invasive safe control policies while also possessing desirable Lyapunov-like qualities such as positive semi-definiteness. We show how these LCBFs can be used to augment a learning-based control policy to guarantee safety and then leverage this approach to develop a safe exploration framework in a MBRL setting. We demonstrate that our approach can handle more general safety constraints than comparative methods via numerical examples.


Full work available at URL: https://arxiv.org/abs/2104.08171




Recommendations




Cites Work


Cited In (22)





This page was built for publication: Safe exploration in model-based reinforcement learning using control barrier functions

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2103658)