Beware of "Explanations" of AI
Beware of "Explanations" of AI
Understanding the decisions made and actions taken by increasingly complex AI system remains a key challenge. This has led to an expanding field of research in explainable artificial intelligence (XAI), highlighting the potential of explanations to enhance trust, support adoption, and meet regulatory standards. However, the question of what constitutes a "good" explanation is dependent on the goals, stakeholders, and context. At a high level, psychological insights such as the concept of mental model alignment can offer guidance, but success in practice is challenging due to social and technical factors. As a result of this ill-defined nature of the problem, explanations can be of poor quality (e.g. unfaithful, irrelevant, or incoherent), potentially leading to substantial risks. Instead of fostering trust and safety, poorly designed explanations can actually cause harm, including wrong decisions, privacy violations, manipulation, and even reduced AI adoption. Therefore, we caution stakeholders to beware of explanations of AI: while they can be vital, they are not automatically a remedy for transparency or responsible AI adoption, and their misuse or limitations can exacerbate harm. Attention to these caveats can help guide future research to improve the quality and impact of AI explanations.
David Martens、Galit Shmueli、Theodoros Evgeniou、Kevin Bauer、Christian Janiesch、Stefan Feuerriegel、Sebastian Gabel、Sofie Goethals、Travis Greene、Nadja Klein、Mathias Kraus、Niklas Kühl、Claudia Perlich、Wouter Verbeke、Alona Zharova、Patrick Zschech、Foster Provost
计算技术、计算机技术
David Martens,Galit Shmueli,Theodoros Evgeniou,Kevin Bauer,Christian Janiesch,Stefan Feuerriegel,Sebastian Gabel,Sofie Goethals,Travis Greene,Nadja Klein,Mathias Kraus,Niklas Kühl,Claudia Perlich,Wouter Verbeke,Alona Zharova,Patrick Zschech,Foster Provost.Beware of "Explanations" of AI[EB/OL].(2025-04-09)[2025-04-27].https://arxiv.org/abs/2504.06791.点此复制
评论