Continual Learning on a Diet: Learning from Sparsely Labeled Streams Under Constrained Computation (2404.12766v2)
Abstract: We propose and study a realistic Continual Learning (CL) setting where learning algorithms are granted a restricted computational budget per time step while training. We apply this setting to large-scale semi-supervised Continual Learning scenarios with sparse label rates. Previous proficient CL methods perform very poorly in this challenging setting. Overfitting to the sparse labeled data and insufficient computational budget are the two main culprits for such a poor performance. Our new setting encourages learning methods to effectively and efficiently utilize the unlabeled data during training. To that end, we propose a simple but highly effective baseline, DietCL, which utilizes both unlabeled and labeled data jointly. DietCL meticulously allocates computational budget for both types of data. We validate our baseline, at scale, on several datasets, e.g., CLOC, ImageNet10K, and CGLM, under constraint budget setups. DietCL outperforms, by a large margin, all existing supervised CL algorithms as well as more recent continual semi-supervised methods. Our extensive analysis and ablations demonstrate that DietCL is stable under a full spectrum of label sparsity, computational budget, and various other ablations.
- Memory aware synapses: Learning what (not) to forget. In European Conference on Computer Vision (ECCV), 2018.
- Task-free continual learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- Continual semi-supervised learning through contrastive interpolation consistency. Pattern Recognition Letters, 2022.
- Special: Self-supervised pretraining for continual learning. Semi-Supervised Learning, pp. 91, 2022.
- New insights on reducing abrupt representation change in online continual learning. In International Conference on Learning Representations (ICLR), 2022.
- Online continual learning with natural distribution shifts: An empirical study with visual data. In International Conference on Computer Vision (ICCV), 2021.
- Efficient lifelong learning with a-gem. In International Conference on Learning Representations (ICLR), 2019a.
- On tiny episodic memories in continual learning. In MTLRL Workshop, International Conference on Machine Learning (ICMLW), 2019b.
- Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition (CVPR), 2009.
- An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations (ICLR), 2020.
- Dytox: Transformers for continual learning with dynamic token expansion. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
- Adversarial continual learning. In European Conference on Computer Vision (ECCV), 2020.
- Self-supervised models are continual learners. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
- Real-time evaluation in online continual learning: A new hope. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11888–11897, 2023.
- Continually learning self-supervised representations with projected functional regularization. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2022.
- Masked autoencoders are scalable vision learners. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
- How well does self-supervised pre-training perform with streaming data? In International Conference on Learning Representations (ICLR), 2021.
- Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 2017.
- Online continual learning on class incremental blurry task configuration with anytime inference. In International Conference on Learning Representations (ICLR), 2021.
- Continual evaluation for lifelong learning: Identifying the stability gap. In International Conference on Learning Representations (ICLR), 2023.
- Avalanche: an end-to-end library for continual learning. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2nd Continual Learning in Computer Vision Workshop, 2021.
- Dualnet: Continual learning, fast and slow. Advances in Neural Information Processing Systems (NeurIPs), 2021.
- Gdumb: A simple approach that questions our progress in continual learning. In European Conference on Computer Vision (ECCV), August 2020.
- Computationally budgeted continual learning: What does matter? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3698–3707, 2023.
- Imagenet-21k pretraining for the masses. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1), 2021.
- Snap. Quarterly results of Snap Inc. Q4, 2017. https://s25.q4cdn.com/442043304/files/doc_presentations/presentation/q4-17-earnings-slides.pdf, 2017. Accessed: 2022-11-06.
- A comprehensive survey of continual learning: Theory, method and application. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2024.
- Learning to prompt for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 139–149, 2022.
- Wenxuan Zhang (75 papers)
- Youssef Mohamed (14 papers)
- Bernard Ghanem (256 papers)
- Philip H. S. Torr (219 papers)
- Adel Bibi (53 papers)
- Mohamed Elhoseiny (102 papers)