Failing with Grace: Learning Neural Network Controllers that are Boundedly Unsafe

Panagiotis Vlantis, Leila Bridgeman, Michael Zavlanos
Proceedings of The 5th Annual Learning for Dynamics and Control Conference, PMLR 211:954-965, 2023.

Abstract

This work considers the problem of learning a feed-forward neural network controller to safely steer an arbitrarily shaped planar robot in a compact, obstacle-occluded workspace. When training neural network controllers, existing closed-loop safety assurances impose stringent data density requirements close to the boundary of the safe state space, which are hard to satisfy in practice. We propose an approach that lifts these strong assumptions and instead admits graceful safety violations, i.e., of a bounded, spatially controlled magnitude. The method employs reachability analysis techniques to include safety constraints in the training process. The method can simultaneously learn a safe vector field for the closed-loop system and provide proven numerical worst-case bounds on safety violations over the whole configuration space, defined by the overlap between an over-approximation of the closed-loop system’s forward reachable set and the set of unsafe states.

Cite this Paper


BibTeX
@InProceedings{pmlr-v211-vlantis23a, title = {Failing with Grace: Learning Neural Network Controllers that are Boundedly Unsafe}, author = {Vlantis, Panagiotis and Bridgeman, Leila and Zavlanos, Michael}, booktitle = {Proceedings of The 5th Annual Learning for Dynamics and Control Conference}, pages = {954--965}, year = {2023}, editor = {Matni, Nikolai and Morari, Manfred and Pappas, George J.}, volume = {211}, series = {Proceedings of Machine Learning Research}, month = {15--16 Jun}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v211/vlantis23a/vlantis23a.pdf}, url = {https://proceedings.mlr.press/v211/vlantis23a.html}, abstract = {This work considers the problem of learning a feed-forward neural network controller to safely steer an arbitrarily shaped planar robot in a compact, obstacle-occluded workspace. When training neural network controllers, existing closed-loop safety assurances impose stringent data density requirements close to the boundary of the safe state space, which are hard to satisfy in practice. We propose an approach that lifts these strong assumptions and instead admits graceful safety violations, i.e., of a bounded, spatially controlled magnitude. The method employs reachability analysis techniques to include safety constraints in the training process. The method can simultaneously learn a safe vector field for the closed-loop system and provide proven numerical worst-case bounds on safety violations over the whole configuration space, defined by the overlap between an over-approximation of the closed-loop system’s forward reachable set and the set of unsafe states.} }
Endnote
%0 Conference Paper %T Failing with Grace: Learning Neural Network Controllers that are Boundedly Unsafe %A Panagiotis Vlantis %A Leila Bridgeman %A Michael Zavlanos %B Proceedings of The 5th Annual Learning for Dynamics and Control Conference %C Proceedings of Machine Learning Research %D 2023 %E Nikolai Matni %E Manfred Morari %E George J. Pappas %F pmlr-v211-vlantis23a %I PMLR %P 954--965 %U https://proceedings.mlr.press/v211/vlantis23a.html %V 211 %X This work considers the problem of learning a feed-forward neural network controller to safely steer an arbitrarily shaped planar robot in a compact, obstacle-occluded workspace. When training neural network controllers, existing closed-loop safety assurances impose stringent data density requirements close to the boundary of the safe state space, which are hard to satisfy in practice. We propose an approach that lifts these strong assumptions and instead admits graceful safety violations, i.e., of a bounded, spatially controlled magnitude. The method employs reachability analysis techniques to include safety constraints in the training process. The method can simultaneously learn a safe vector field for the closed-loop system and provide proven numerical worst-case bounds on safety violations over the whole configuration space, defined by the overlap between an over-approximation of the closed-loop system’s forward reachable set and the set of unsafe states.
APA
Vlantis, P., Bridgeman, L. & Zavlanos, M.. (2023). Failing with Grace: Learning Neural Network Controllers that are Boundedly Unsafe. Proceedings of The 5th Annual Learning for Dynamics and Control Conference, in Proceedings of Machine Learning Research 211:954-965 Available from https://proceedings.mlr.press/v211/vlantis23a.html.

Related Material