ROAD: Responsibility-Oriented Reward Design for Reinforcement Learning in Autonomous Driving
ROAD: Responsibility-Oriented Reward Design for Reinforcement Learning in Autonomous Driving
Reinforcement learning (RL) in autonomous driving employs a trial-and-error mechanism, enhancing robustness in unpredictable environments. However, crafting effective reward functions remains challenging, as conventional approaches rely heavily on manual design and demonstrate limited efficacy in complex scenarios. To address this issue, this study introduces a responsibility-oriented reward function that explicitly incorporates traffic regulations into the RL framework. Specifically, we introduced a Traffic Regulation Knowledge Graph and leveraged Vision-Language Models alongside Retrieval-Augmented Generation techniques to automate reward assignment. This integration guides agents to adhere strictly to traffic laws, thus minimizing rule violations and optimizing decision-making performance in diverse driving conditions. Experimental validations demonstrate that the proposed methodology significantly improves the accuracy of assigning accident responsibilities and effectively reduces the agent's liability in traffic incidents.
Yongming Chen、Miner Chen、Liewen Liao、Mingyang Jiang、Xiang Zuo、Hengrui Zhang、Yuchen Xi、Songan Zhang
公路运输工程
Yongming Chen,Miner Chen,Liewen Liao,Mingyang Jiang,Xiang Zuo,Hengrui Zhang,Yuchen Xi,Songan Zhang.ROAD: Responsibility-Oriented Reward Design for Reinforcement Learning in Autonomous Driving[EB/OL].(2025-05-30)[2025-07-22].https://arxiv.org/abs/2505.24317.点此复制
评论