|国家预印本平台
首页|Memory-Statistics Tradeoff in Continual Learning with Structural Regularization

Memory-Statistics Tradeoff in Continual Learning with Structural Regularization

Memory-Statistics Tradeoff in Continual Learning with Structural Regularization

来源:Arxiv_logoArxiv
英文摘要

We study the statistical performance of a continual learning problem with two linear regression tasks in a well-specified random design setting. We consider a structural regularization algorithm that incorporates a generalized $\ell_2$-regularization tailored to the Hessian of the previous task for mitigating catastrophic forgetting. We establish upper and lower bounds on the joint excess risk for this algorithm. Our analysis reveals a fundamental trade-off between memory complexity and statistical efficiency, where memory complexity is measured by the number of vectors needed to define the structural regularization. Specifically, increasing the number of vectors in structural regularization leads to a worse memory complexity but an improved excess risk, and vice versa. Furthermore, our theory suggests that naive continual learning without regularization suffers from catastrophic forgetting, while structural regularization mitigates this issue. Notably, structural regularization achieves comparable performance to joint training with access to both tasks simultaneously. These results highlight the critical role of curvature-aware regularization for continual learning.

Haoran Li、Jingfeng Wu、Vladimir Braverman

计算技术、计算机技术

Haoran Li,Jingfeng Wu,Vladimir Braverman.Memory-Statistics Tradeoff in Continual Learning with Structural Regularization[EB/OL].(2025-04-04)[2025-05-21].https://arxiv.org/abs/2504.04039.点此复制

评论