A Test-Function Approach to Incremental Stability
A Test-Function Approach to Incremental Stability
This paper presents a novel framework for analyzing Incremental-Input-to-State Stability ($δ$ISS) based on the idea of using rewards as "test functions." Whereas control theory traditionally deals with Lyapunov functions that satisfy a time-decrease condition, reinforcement learning (RL) value functions are constructed by exponentially decaying a Lipschitz reward function that may be non-smooth and unbounded on both sides. Thus, these RL-style value functions cannot be directly understood as Lyapunov certificates. We develop a new equivalence between a variant of incremental input-to-state stability of a closed-loop system under given a policy, and the regularity of RL-style value functions under adversarial selection of a Hölder-continuous reward function. This result highlights that the regularity of value functions, and their connection to incremental stability, can be understood in a way that is distinct from the traditional Lyapunov-based approach to certifying stability in control theory.
Daniel Pfrommer、Max Simchowitz、Ali Jadbabaie
自动化基础理论计算技术、计算机技术
Daniel Pfrommer,Max Simchowitz,Ali Jadbabaie.A Test-Function Approach to Incremental Stability[EB/OL].(2025-07-01)[2025-08-02].https://arxiv.org/abs/2507.00695.点此复制
评论