An Adaptive Method Stabilizing Activations for Enhanced Generalization
An Adaptive Method Stabilizing Activations for Enhanced Generalization
We introduce AdaAct, a novel optimization algorithm that adjusts learning rates according to activation variance. Our method enhances the stability of neuron outputs by incorporating neuron-wise adaptivity during the training process, which subsequently leads to better generalization -- a complementary approach to conventional activation regularization methods. Experimental results demonstrate AdaAct's competitive performance across standard image classification benchmarks. We evaluate AdaAct on CIFAR and ImageNet, comparing it with other state-of-the-art methods. Importantly, AdaAct effectively bridges the gap between the convergence speed of Adam and the strong generalization capabilities of SGD, all while maintaining competitive execution times. Code is available at https://github.com/hseung88/adaact.
Hyunseok Seung、Jaewoo Lee、Hyunsuk Ko
计算技术、计算机技术
Hyunseok Seung,Jaewoo Lee,Hyunsuk Ko.An Adaptive Method Stabilizing Activations for Enhanced Generalization[EB/OL].(2025-06-09)[2025-07-16].https://arxiv.org/abs/2506.08353.点此复制
评论