Online Continual Learning via Spiking Neural Networks with Sleep Enhanced Latent Replay
Online Continual Learning via Spiking Neural Networks with Sleep Enhanced Latent Replay
Edge computing scenarios necessitate the development of hardware-efficient online continual learning algorithms to be adaptive to dynamic environment. However, existing algorithms always suffer from high memory overhead and bias towards recently trained tasks. To tackle these issues, this paper proposes a novel online continual learning approach termed as SESLR, which incorporates a sleep enhanced latent replay scheme with spiking neural networks (SNNs). SESLR leverages SNNs' binary spike characteristics to store replay features in single bits, significantly reducing memory overhead. Furthermore, inspired by biological sleep-wake cycles, SESLR introduces a noise-enhanced sleep phase where the model exclusively trains on replay samples with controlled noise injection, effectively mitigating classification bias towards new classes. Extensive experiments on both conventional (MNIST, CIFAR10) and neuromorphic (NMNIST, CIFAR10-DVS) datasets demonstrate SESLR's effectiveness. On Split CIFAR10, SESLR achieves nearly 30% improvement in average accuracy with only one-third of the memory consumption compared to baseline methods. On Split CIFAR10-DVS, it improves accuracy by approximately 10% while reducing memory overhead by a factor of 32. These results validate SESLR as a promising solution for online continual learning in resource-constrained edge computing scenarios.
Erliang Lin、Wenbin Luo、Wei Jia、Yu Chen、Shaofu Yang
计算技术、计算机技术
Erliang Lin,Wenbin Luo,Wei Jia,Yu Chen,Shaofu Yang.Online Continual Learning via Spiking Neural Networks with Sleep Enhanced Latent Replay[EB/OL].(2025-07-10)[2025-07-16].https://arxiv.org/abs/2507.02901.点此复制
评论