Stable modular control via contraction theory for reinforcement learning

Bing Song, Jean-Jacques Slotine, Quang-Cuong Pham
Proceedings of the 6th Annual Learning for Dynamics & Control Conference, PMLR 242:1136-1148, 2024.

Abstract

We propose a novel way to integrate control theoretical results with reinforcement learning (RL) for stability, robustness, and generalization: developing modular control architectures via contraction theory to simplify the complex problems. To guarantee control stability for RL, we leverage modularity to deconstruct the nonlinear stability problems into algebraically solvable ones, yielding linear constraints on the input gradients of control networks that can be as simple as switching the signs of network weights. This control architecture can be implemented in general RL frameworks without modifying the algorithms. This minimally invasive way allows arguably easy integration into hierarchical RL, and improves its performance. We realize the modularity by constructing an auxiliary space through coordinate transformation. Within the auxiliary space, system dynamics can be represented as hierarchical combinations of subsystems. These subsystems converge recursively following their hierarchies, provided stable self-feedbacks. We implement this modular control architecture in PPO and hierarchical RL, and demonstrate in simulation (i) the necessity of control stability for robustness and generalization and (ii) the effectiveness in improving hierarchical RL for manipulation learning.

Cite this Paper


BibTeX
@InProceedings{pmlr-v242-song24a, title = {Stable modular control via contraction theory for reinforcement learning}, author = {Song, Bing and Slotine, Jean-Jacques and Pham, Quang-Cuong}, booktitle = {Proceedings of the 6th Annual Learning for Dynamics & Control Conference}, pages = {1136--1148}, year = {2024}, editor = {Abate, Alessandro and Cannon, Mark and Margellos, Kostas and Papachristodoulou, Antonis}, volume = {242}, series = {Proceedings of Machine Learning Research}, month = {15--17 Jul}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v242/song24a/song24a.pdf}, url = {https://proceedings.mlr.press/v242/song24a.html}, abstract = {We propose a novel way to integrate control theoretical results with reinforcement learning (RL) for stability, robustness, and generalization: developing modular control architectures via contraction theory to simplify the complex problems. To guarantee control stability for RL, we leverage modularity to deconstruct the nonlinear stability problems into algebraically solvable ones, yielding linear constraints on the input gradients of control networks that can be as simple as switching the signs of network weights. This control architecture can be implemented in general RL frameworks without modifying the algorithms. This minimally invasive way allows arguably easy integration into hierarchical RL, and improves its performance. We realize the modularity by constructing an auxiliary space through coordinate transformation. Within the auxiliary space, system dynamics can be represented as hierarchical combinations of subsystems. These subsystems converge recursively following their hierarchies, provided stable self-feedbacks. We implement this modular control architecture in PPO and hierarchical RL, and demonstrate in simulation (i) the necessity of control stability for robustness and generalization and (ii) the effectiveness in improving hierarchical RL for manipulation learning.} }
Endnote
%0 Conference Paper %T Stable modular control via contraction theory for reinforcement learning %A Bing Song %A Jean-Jacques Slotine %A Quang-Cuong Pham %B Proceedings of the 6th Annual Learning for Dynamics & Control Conference %C Proceedings of Machine Learning Research %D 2024 %E Alessandro Abate %E Mark Cannon %E Kostas Margellos %E Antonis Papachristodoulou %F pmlr-v242-song24a %I PMLR %P 1136--1148 %U https://proceedings.mlr.press/v242/song24a.html %V 242 %X We propose a novel way to integrate control theoretical results with reinforcement learning (RL) for stability, robustness, and generalization: developing modular control architectures via contraction theory to simplify the complex problems. To guarantee control stability for RL, we leverage modularity to deconstruct the nonlinear stability problems into algebraically solvable ones, yielding linear constraints on the input gradients of control networks that can be as simple as switching the signs of network weights. This control architecture can be implemented in general RL frameworks without modifying the algorithms. This minimally invasive way allows arguably easy integration into hierarchical RL, and improves its performance. We realize the modularity by constructing an auxiliary space through coordinate transformation. Within the auxiliary space, system dynamics can be represented as hierarchical combinations of subsystems. These subsystems converge recursively following their hierarchies, provided stable self-feedbacks. We implement this modular control architecture in PPO and hierarchical RL, and demonstrate in simulation (i) the necessity of control stability for robustness and generalization and (ii) the effectiveness in improving hierarchical RL for manipulation learning.
APA
Song, B., Slotine, J. & Pham, Q.. (2024). Stable modular control via contraction theory for reinforcement learning. Proceedings of the 6th Annual Learning for Dynamics & Control Conference, in Proceedings of Machine Learning Research 242:1136-1148 Available from https://proceedings.mlr.press/v242/song24a.html.

Related Material