Calibration Attention: Instance-wise Temperature Scaling for Vision Transformers
Calibration Attention: Instance-wise Temperature Scaling for Vision Transformers
Probability calibration is critical when Vision Transformers are deployed in risk-sensitive applications. The standard fix, post-hoc temperature scaling, uses a single global scalar and requires a held-out validation set. We introduce Calibration Attention (CalAttn), a drop-in module that learns an adaptive, per-instance temperature directly from the ViT's CLS token. Across CIFAR-10/100, MNIST, Tiny-ImageNet, and ImageNet-1K, CalAttn reduces calibration error by up to 4x on ViT-224, DeiT, and Swin, while adding under 0.1 percent additional parameters. The learned temperatures cluster tightly around 1.0, in contrast to the large global values used by standard temperature scaling. CalAttn is simple, efficient, and architecture-agnostic, and yields more trustworthy probabilities without sacrificing accuracy. Code: [https://github.com/EagleAdelaide/CalibrationAttention-CalAttn-](https://github.com/EagleAdelaide/CalibrationAttention-CalAttn-)
Wenhao Liang、Wei Emma Zhang、Lin Yue、Miao Xu、Olaf Maennel、Weitong Chen
计算技术、计算机技术
Wenhao Liang,Wei Emma Zhang,Lin Yue,Miao Xu,Olaf Maennel,Weitong Chen.Calibration Attention: Instance-wise Temperature Scaling for Vision Transformers[EB/OL].(2025-08-12)[2025-08-24].https://arxiv.org/abs/2508.08547.点此复制
评论