|国家预印本平台
首页|Soft-CAM: Making black box models self-explainable for high-stakes decisions

Soft-CAM: Making black box models self-explainable for high-stakes decisions

Soft-CAM: Making black box models self-explainable for high-stakes decisions

来源:Arxiv_logoArxiv
英文摘要

Convolutional neural networks (CNNs) are widely used for high-stakes applications like medicine, often surpassing human performance. However, most explanation methods rely on post-hoc attribution, approximating the decision-making process of already trained black-box models. These methods are often sensitive, unreliable, and fail to reflect true model reasoning, limiting their trustworthiness in critical applications. In this work, we introduce SoftCAM, a straightforward yet effective approach that makes standard CNN architectures inherently interpretable. By removing the global average pooling layer and replacing the fully connected classification layer with a convolution-based class evidence layer, SoftCAM preserves spatial information and produces explicit class activation maps that form the basis of the model's predictions. Evaluated on three medical datasets, SoftCAM maintains classification performance while significantly improving both the qualitative and quantitative explanation compared to existing post-hoc methods. Our results demonstrate that CNNs can be inherently interpretable without compromising performance, advancing the development of self-explainable deep learning for high-stakes decision-making.

Kerol Djoumessi、Philipp Berens

医学现状、医学发展医学研究方法

Kerol Djoumessi,Philipp Berens.Soft-CAM: Making black box models self-explainable for high-stakes decisions[EB/OL].(2025-05-23)[2025-07-25].https://arxiv.org/abs/2505.17748.点此复制

评论