A Numerically Efficient Method to Enhance Model Predictive Control Performance with a Reinforcement Learning Policy
A Numerically Efficient Method to Enhance Model Predictive Control Performance with a Reinforcement Learning Policy
We propose a novel approach for combining model predictive control (MPC) with reinforcement learning (RL) to reduce online computation while achieving high closed-loop tracking performance and constraint satisfaction. This method, called Policy-Enhanced Partial Tightening (PEPT), approximates the optimal value function through a Riccati recursion around a state-control trajectory obtained by evaluating the RL policy. The result is a convex quadratic terminal cost that can be seamlessly integrated into the MPC formulation. The proposed controller is tested in simulations on a trajectory tracking problem for a quadcopter with nonlinear dynamics and bounded state and control. The results highlight PEPT's effectiveness, outperforming both pure RL policies and several MPC variations. Compared to pure RL, PEPT achieves 1000 times lower constraint violation cost with only twice the feedback time. Against the best MPC-based policy, PEPT reduces constraint violations by 2 to 5 times and runs nearly 3 times faster while maintaining similar tracking performance. The code is open-source at www.github.com/aghezz1/pept.
Andrea Ghezzi、Rudolf Reiter、Katrin Baumg?rtner、Alberto Bemporad、Moritz Diehl
自动化技术、自动化技术设备计算技术、计算机技术
Andrea Ghezzi,Rudolf Reiter,Katrin Baumg?rtner,Alberto Bemporad,Moritz Diehl.A Numerically Efficient Method to Enhance Model Predictive Control Performance with a Reinforcement Learning Policy[EB/OL].(2025-04-03)[2025-05-04].https://arxiv.org/abs/2504.02710.点此复制
评论