|国家预印本平台
首页|RCR-AF: Enhancing Model Generalization via Rademacher Complexity Reduction Activation Function

RCR-AF: Enhancing Model Generalization via Rademacher Complexity Reduction Activation Function

RCR-AF: Enhancing Model Generalization via Rademacher Complexity Reduction Activation Function

来源:Arxiv_logoArxiv
英文摘要

Despite their widespread success, deep neural networks remain critically vulnerable to adversarial attacks, posing significant risks in safety-sensitive applications. This paper investigates activation functions as a crucial yet underexplored component for enhancing model robustness. We propose a Rademacher Complexity Reduction Activation Function (RCR-AF), a novel activation function designed to improve both generalization and adversarial resilience. RCR-AF uniquely combines the advantages of GELU (including smoothness, gradient stability, and negative information retention) with ReLU's desirable monotonicity, while simultaneously controlling both model sparsity and capacity through built-in clipping mechanisms governed by two hyperparameters, $α$ and $γ$. Our theoretical analysis, grounded in Rademacher complexity, demonstrates that these parameters directly modulate the model's Rademacher complexity, offering a principled approach to enhance robustness. Comprehensive empirical evaluations show that RCR-AF consistently outperforms widely-used alternatives (ReLU, GELU, and Swish) in both clean accuracy under standard training and in adversarial robustness within adversarial training paradigms.

Yunrui Yu、Kafeng Wang、Hang Su、Jun Zhu

计算技术、计算机技术

Yunrui Yu,Kafeng Wang,Hang Su,Jun Zhu.RCR-AF: Enhancing Model Generalization via Rademacher Complexity Reduction Activation Function[EB/OL].(2025-07-30)[2025-08-06].https://arxiv.org/abs/2507.22446.点此复制

评论