🤖 AI Summary
To address the dual challenges of high memory overhead and catastrophic forgetting in memory-based continual learning, this paper proposes a training-free, lightweight generator based on singular value decomposition (SVD). The generator models class-wise data distributions efficiently by fitting only a small number of samples in linear time, drastically reducing memory footprint. Integrated into both A-GEM and experience replay frameworks, it enables low-overhead, high-fidelity knowledge retention without parameter updates. Extensive experiments across standard benchmarks—including CIFAR-100 and ImageNet-Subset—and diverse architectures—ResNet and ViT—demonstrate that our method reduces memory consumption by 30–60% while improving average accuracy over baselines by 2.1–5.7 percentage points. These results validate its superior trade-off between memory efficiency and generalization capability in continual learning settings.
📝 Abstract
Catastrophic forgetting can be trivially alleviated by keeping all data from previous tasks in memory. Therefore, minimizing the memory footprint while maximizing the amount of relevant information is crucial to the challenge of continual learning. This paper aims to decrease required memory for memory-based continuous learning algorithms. We explore the options of extracting a minimal amount of information, while maximally alleviating forgetting. We propose the usage of lightweight generators based on Singular Value Decomposition to enhance existing continual learning methods, such as A-GEM and Experience Replay. These generators need a minimal amount of memory while being maximally effective. They require no training time, just a single linear-time fitting step, and can capture a distribution effectively from a small number of data samples. Depending on the dataset and network architecture, our results show a significant increase in average accuracy compared to the original methods. Our method shows great potential in minimizing the memory footprint of memory-based continual learning algorithms.