|国家预印本平台
首页|A Physics-Inspired Optimizer: Velocity Regularized Adam

A Physics-Inspired Optimizer: Velocity Regularized Adam

A Physics-Inspired Optimizer: Velocity Regularized Adam

来源:Arxiv_logoArxiv
英文摘要

We introduce Velocity-Regularized Adam (VRAdam), a physics-inspired optimizer for training deep neural networks that draws on ideas from quartic terms for kinetic energy with its stabilizing effects on various system dynamics. Previous algorithms, including the ubiquitous Adam, operate at the so called adaptive edge of stability regime during training leading to rapid oscillations and slowed convergence of loss. However, VRAdam adds a higher order penalty on the learning rate based on the velocity such that the algorithm automatically slows down whenever weight updates become large. In practice, we observe that the effective dynamic learning rate shrinks in high-velocity regimes, damping oscillations and allowing for a more aggressive base step size when necessary without divergence. By combining this velocity-based regularizer for global damping with per-parameter scaling of Adam to create a hybrid optimizer, we demonstrate that VRAdam consistently exceeds the performance against standard optimizers including AdamW. We benchmark various tasks such as image classification, language modeling, image generation and generative modeling using diverse architectures and training methodologies including Convolutional Neural Networks (CNNs), Transformers, and GFlowNets.

Pranav Vaidhyanathan、Lucas Schorling、Natalia Ares、Michael A. Osborne

计算技术、计算机技术

Pranav Vaidhyanathan,Lucas Schorling,Natalia Ares,Michael A. Osborne.A Physics-Inspired Optimizer: Velocity Regularized Adam[EB/OL].(2025-05-19)[2025-06-13].https://arxiv.org/abs/2505.13196.点此复制

评论