Random Representations Outperform Online Continually Learned Representations (2402.08823v3)
Abstract: Continual learning has primarily focused on the issue of catastrophic forgetting and the associated stability-plasticity tradeoffs. However, little attention has been paid to the efficacy of continually learned representations, as representations are learned alongside classifiers throughout the learning process. Our primary contribution is empirically demonstrating that existing online continually trained deep networks produce inferior representations compared to a simple pre-defined random transforms. Our approach projects raw pixels using a fixed random transform, approximating an RBF-Kernel initialized before any data is seen. We then train a simple linear classifier on top without storing any exemplars, processing one sample at a time in an online continual learning setting. This method, called RanDumb, significantly outperforms state-of-the-art continually learned representations across all standard online continual learning benchmarks. Our study reveals the significant limitations of representation learning, particularly in low-exemplar and online continual learning scenarios. Extending our investigation to popular exemplar-free scenarios with pretrained models, we find that training only a linear classifier on top of pretrained representations surpasses most continual fine-tuning and prompt-tuning strategies. Overall, our investigation challenges the prevailing assumptions about effective representation learning in online continual learning. Our code is available at://github.com/drimpossible/RanDumb.
- Ss-il: Separated softmax for incremental learning. In ICCV, 2021.
- Online continual learning with maximally interfered retrieval. In NeurIPS, 2019a.
- Gradient based sample selection for online continual learning. In NeurIPS, 2019b.
- Anonymous. Meta continual learning revisited: Implicitly enhancing online hessian approximation via variance reduction. In ICLR, 2024.
- Learning fast, learning slow: A general continual learning method based on complementary learning system. In ICLR, 2022.
- Rainbow memory: Continual learning with a memory of diverse samples. In CVPR, 2021.
- Edges are the’independent components’ of natural scenes. In NeurIPS, 1996.
- Class-incremental continual learning into the extended der-verse. TPAMI, 2022.
- Dark experience for general continual learning: a strong, simple baseline. In NeurIPS, 2020.
- New insights on reducing abrupt representation change in online continual learning. In ICLR, 2022.
- Co2l: Contrastive continual learning. In ICCV, 2021.
- Efficient lifelong learning with a-gem. In ICLR, 2019a.
- Continual learning with tiny episodic memories. In ICML-W, 2019b.
- Online bias correction for task-free continual learning. In ICLR, 2023.
- Continual prototype evolution: Learning online from non-stationary data streams. In ICCV, 2021.
- Continual learning: A comparative study on how to defy forgetting in classification tasks. In TPAMI, 2020.
- Tic-clip: Continual training of clip models. ArXiv, 2023.
- Online continual learning through mutual information maximization. In ICML, 2022.
- Look-ahead meta learning for continual learning. NeurIPS, 2020.
- Lifelong machine learning with deep streaming linear discriminant analysis. In CVPR-W, 2020.
- Knowledge accumulation in continually learned representations and the issue of feature forgetting. arXiv preprint arXiv:2304.00933, 2023.
- A simple baseline that questions the use of pretrained-models in continual learning. In NeurIPS-W, 2022.
- Gradient based memory editing for task-free continual learning. In NeurIPS, 2021.
- Overcoming catastrophic forgetting in neural networks. PNAS, 2017.
- A neural dirichlet process mixture model for task-free continual learning. In ICLR, 2020.
- Learning without forgetting. TPAMI, 2017.
- Gradient episodic memory for continual learning. In NeurIPS, 2017.
- Supervised contrastive replay: Revisiting the nearest class mean classifier in online class-incremental continual learning. In CVPR, 2021.
- Online continual learning in image classification: An empirical survey. Neurocomputing, 2022.
- Ranpac: Random projections and pre-trained models for continual learning. In NeurIPS, 2023.
- Goeffrey J McLachlan. Mahalanobis distance. Resonance, 4(6):20–26, 1999.
- Geoffrey J McLachlan. Discriminant analysis and statistical pattern recognition. John Wiley & Sons, 2005.
- Distance-based image classification: Generalizing to new classes at near-zero cost. TPAMI, 2013.
- Architecture matters in continual learning. arXiv preprint arXiv:2202.00275, 2022.
- Essentials for class incremental learning. In CVPR, 2021.
- Fine-tuning can cripple your foundation model; preserving features may be the solution. arXiv preprint arXiv:2308.13320, 2023.
- A wholistic view of continual learning with deep neural networks: Forgotten lessons and the bridge to active and open world learning. Neural Networks, 2023.
- A review of kernel methods for feature extraction in nonlinear process monitoring. Processes, 8(1), 2020.
- Gdumb: A simple approach that questions our progress in continual learning. In ECCV, 2020.
- Online continual learning without the storage constraint. arXiv preprint arXiv:2305.09253, 2023a.
- Computationally budgeted continual learning: What does matter? In CVPR, 2023b.
- Random features for large-scale kernel machines. NeurIPS, 2007.
- Continual unsupervised representation learning. NeurIPS, 32, 2019.
- icarl: Incremental classifier and representation learning. In CVPR, 2017.
- Learning to learn without forgetting by maximizing transfer and minimizing interference. In ICLR, 2019.
- Online class-incremental continual learning with adversarial shapley value. In AAAI, 2021.
- Coda-prompt: Continual decomposed attention-based prompting for rehearsal-free continual learning. In CVPR, 2023a.
- A closer look at rehearsal-free continual learning. In CVPR-W, 2023b.
- Gido M van de Ven and Andreas S Tolias. Three scenarios for continual learning. In NeurIPS-W, 2018.
- Class-incremental learning with generative classifiers. In CVPR-W, 2021.
- Continual learning: Applications and the road forward. arXiv preprint arXiv:2311.11908, 2023.
- Foster: Feature boosting and compression for class-incremental learning. In ECCV, 2022a.
- Dualprompt: Complementary prompting for rehearsal-free continual learning. In European Conference on Computer Vision (ECCV), 2022b.
- Learning to prompt for continual learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022c.
- Online prototype learning for online continual learning. In ICCV, 2023.
- Large scale incremental learning. In CVPR, 2019.
- Continual variational autoencoder learning via online cooperative memorization. In ECCV, 2022.
- Self-evolved dynamic expansion model for task-free continual learning. In ICCV, 2023.
- Prediction error-based classification for class-incremental learning. ICLR, 2024.
- Continual learning through synaptic intelligence. In ICML, 2017.
- Task agnostic continual learning using online variational bayes. arXiv preprint arXiv:1803.10123, 2018.
- Slca: Slow learner with classifier alignment for continual learning on a pre-trained model. In ICCV, 2023.
- A model or 603 exemplars: Towards memory-efficient class-incremental learning. arXiv preprint arXiv:2205.13218, 2022.
- Revisiting class-incremental learning with pre-trained models: Generalizability and adaptivity are all you need. arXiv preprint arXiv:2303.07338, 2023.
- Class-incremental learning via dual augmentation. NeurIPS, 2021.