Challenges in Model Agnostic Controller Learning for Unstable Systems
Challenges in Model Agnostic Controller Learning for Unstable Systems
Model agnostic controller learning, for instance by direct policy optimization, has been the object of renewed attention lately, since it avoids a computationally expensive system identification step. Indeed, direct policy search has been empirically shown to lead to optimal controllers in a number of cases of practical importance. However, to date, these empirical results have not been backed up with a comprehensive theoretical analysis for general problems. In this paper we use a simple example to show that direct policy optimization is not directly generalizable to other seemingly simple problems. In such cases, direct optimization of a performance index can lead to unstable pole/zero cancellations, resulting in the loss of internal stability and unbounded outputs in response to arbitrarily small perturbations. We conclude the paper by analyzing several alternatives to avoid this phenomenon, suggesting some new directions in direct control policy optimization.
Mario Sznaier、Mustafa Bozdag
自动化基础理论
Mario Sznaier,Mustafa Bozdag.Challenges in Model Agnostic Controller Learning for Unstable Systems[EB/OL].(2025-05-16)[2025-06-07].https://arxiv.org/abs/2505.11641.点此复制
评论