Explainable AI for Safe and Trustworthy Autonomous Driving: A Systematic Review
Explainable AI for Safe and Trustworthy Autonomous Driving: A Systematic Review
Artificial Intelligence (AI) shows promising applications for the perception and planning tasks in autonomous driving (AD) due to its superior performance compared to conventional methods. However, inscrutable AI systems exacerbate the existing challenge of safety assurance of AD. One way to mitigate this challenge is to utilize explainable AI (XAI) techniques. To this end, we present the first comprehensive systematic literature review of explainable methods for safe and trustworthy AD. We begin by analyzing the requirements for AI in the context of AD, focusing on three key aspects: data, model, and agency. We find that XAI is fundamental to meeting these requirements. Based on this, we explain the sources of explanations in AI and describe a taxonomy of XAI. We then identify five key contributions of XAI for safe and trustworthy AI in AD, which are interpretable design, interpretable surrogate models, interpretable monitoring, auxiliary explanations, and interpretable validation. Finally, we propose a modular framework called SafeX to integrate these contributions, enabling explanation delivery to users while simultaneously ensuring the safety of AI models.
Stefano V. Albrecht、Anton Kuznietsov、Cheng Wang、Balint Gyevnar、Steven Peters
自动化技术经济自动化技术、自动化技术设备计算技术、计算机技术
Stefano V. Albrecht,Anton Kuznietsov,Cheng Wang,Balint Gyevnar,Steven Peters.Explainable AI for Safe and Trustworthy Autonomous Driving: A Systematic Review[EB/OL].(2024-02-08)[2025-06-27].https://arxiv.org/abs/2402.10086.点此复制
评论