On the Robustness of Derivative-free Methods for Linear Quadratic Regulator
On the Robustness of Derivative-free Methods for Linear Quadratic Regulator
Policy optimization has drawn increasing attention in reinforcement learning, particularly in the context of derivative-free methods for linear quadratic regulator (LQR) problems with unknown dynamics. This paper focuses on characterizing the robustness of derivative-free methods for solving an infinite-horizon LQR problem. To be specific, we estimate policy gradients by cost values, and study the effect of perturbations on the estimations, where the perturbations may arise from function approximations, measurement noises, etc. We show that under sufficiently small perturbations, the derivative-free methods converge to any pre-specified neighborhood of the optimal policy. Furthermore, we establish explicit bounds on the perturbations, and provide the sample complexity for the perturbed derivative-free methods.
Weijian Li、Panagiotis Kounatidis、Zhong-Ping Jiang、Andreas A. Malikopoulos
自动化基础理论计算技术、计算机技术
Weijian Li,Panagiotis Kounatidis,Zhong-Ping Jiang,Andreas A. Malikopoulos.On the Robustness of Derivative-free Methods for Linear Quadratic Regulator[EB/OL].(2025-06-14)[2025-07-25].https://arxiv.org/abs/2506.12596.点此复制
评论