Escaping Plato's Cave: Robust Conceptual Reasoning through Interpretable 3D Neural Object Volumes
Escaping Plato's Cave: Robust Conceptual Reasoning through Interpretable 3D Neural Object Volumes
With the rise of neural networks, especially in high-stakes applications, these networks need two properties (i) robustness and (ii) interpretability to ensure their safety. Recent advances in classifiers with 3D volumetric object representations have demonstrated a greatly enhanced robustness in out-of-distribution data. However, these 3D-aware classifiers have not been studied from the perspective of interpretability. We introduce CAVE - Concept Aware Volumes for Explanations - a new direction that unifies interpretability and robustness in image classification. We design an inherently-interpretable and robust classifier by extending existing 3D-aware classifiers with concepts extracted from their volumetric representations for classification. In an array of quantitative metrics for interpretability, we compare against different concept-based approaches across the explainable AI literature and show that CAVE discovers well-grounded concepts that are used consistently across images, while achieving superior robustness.
Nhi Pham、Bernt Schiele、Adam Kortylewski、Jonas Fischer
计算技术、计算机技术
Nhi Pham,Bernt Schiele,Adam Kortylewski,Jonas Fischer.Escaping Plato's Cave: Robust Conceptual Reasoning through Interpretable 3D Neural Object Volumes[EB/OL].(2025-03-17)[2025-08-02].https://arxiv.org/abs/2503.13429.点此复制
评论