|国家预印本平台
首页|Your Attention Matters: to Improve Model Robustness to Noise and Spurious Correlations

Your Attention Matters: to Improve Model Robustness to Noise and Spurious Correlations

Your Attention Matters: to Improve Model Robustness to Noise and Spurious Correlations

来源:Arxiv_logoArxiv
英文摘要

Self-attention mechanisms are foundational to Transformer architectures, supporting their impressive success in a wide range of tasks. While there are many self-attention variants, their robustness to noise and spurious correlations has not been well studied. This study evaluates Softmax, Sigmoid, Linear, Doubly Stochastic, and Cosine attention within Vision Transformers under different data corruption scenarios. Through testing across the CIFAR-10, CIFAR-100, and Imagenette datasets, we show that Doubly Stochastic attention is the most robust. It consistently outperformed the next best mechanism by $0.1\%-5.1\%$ when training data, or both training and testing data, were corrupted. Our findings inform self-attention selection in contexts with imperfect data. The code used is available at https://github.com/ctamayor/NeurIPS-Robustness-ViT.

Camilo Tamayo-Rousseau、Yunjia Zhao、Yiqun Zhang、Randall Balestriero

计算技术、计算机技术

Camilo Tamayo-Rousseau,Yunjia Zhao,Yiqun Zhang,Randall Balestriero.Your Attention Matters: to Improve Model Robustness to Noise and Spurious Correlations[EB/OL].(2025-08-04)[2025-08-10].https://arxiv.org/abs/2507.20453.点此复制

评论