Greedy Low-Rank Gradient Compression for Distributed Learning with Convergence Guarantees
Greedy Low-Rank Gradient Compression for Distributed Learning with Convergence Guarantees
Distributed optimization is pivotal for large-scale signal processing and machine learning, yet communication overhead remains a major bottleneck. Low-rank gradient compression, in which the transmitted gradients are approximated by low-rank matrices to reduce communication, offers a promising remedy. Existing methods typically adopt either randomized or greedy compression strategies: randomized approaches project gradients onto randomly chosen subspaces, introducing high variance and degrading empirical performance; greedy methods select the most informative subspaces, achieving strong empirical results but lacking convergence guarantees. To address this gap, we propose GreedyLore--the first Greedy Low-Rank gradient compression algorithm for distributed learning with rigorous convergence guarantees. GreedyLore incorporates error feedback to correct the bias introduced by greedy compression and introduces a semi-lazy subspace update that ensures the compression operator remains contractive throughout all iterations. With these techniques, we prove that GreedyLore achieves a convergence rate of $\mathcal{O}(Ï/\sqrt{NT} + 1/T)$ under standard optimizers such as MSGD and Adam--marking the first linear speedup convergence rate for low-rank gradient compression. Extensive experiments are conducted to validate our theoretical findings.
Chuyan Chen、Yutong He、Pengrui Li、Weichen Jia、Kun Yuan
计算技术、计算机技术
Chuyan Chen,Yutong He,Pengrui Li,Weichen Jia,Kun Yuan.Greedy Low-Rank Gradient Compression for Distributed Learning with Convergence Guarantees[EB/OL].(2025-07-11)[2025-07-25].https://arxiv.org/abs/2507.08784.点此复制
评论