Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Realistic Continual Learning Approach using Pre-trained Models (2404.07729v1)

Published 11 Apr 2024 in cs.LG and cs.CV

Abstract: Continual learning (CL) is crucial for evaluating adaptability in learning solutions to retain knowledge. Our research addresses the challenge of catastrophic forgetting, where models lose proficiency in previously learned tasks as they acquire new ones. While numerous solutions have been proposed, existing experimental setups often rely on idealized class-incremental learning scenarios. We introduce Realistic Continual Learning (RealCL), a novel CL paradigm where class distributions across tasks are random, departing from structured setups. We also present CLARE (Continual Learning Approach with pRE-trained models for RealCL scenarios), a pre-trained model-based solution designed to integrate new knowledge while preserving past learning. Our contributions include pioneering RealCL as a generalization of traditional CL setups, proposing CLARE as an adaptable approach for RealCL tasks, and conducting extensive experiments demonstrating its effectiveness across various RealCL scenarios. Notably, CLARE outperforms existing models on RealCL benchmarks, highlighting its versatility and robustness in unpredictable learning environments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (48)
  1. Gradient based sample selection for online continual learning, in: Neural Information Processing Systems.
  2. Entropy-based stability-plasticity for lifelong learning. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) , 3720–3727.
  3. Rethinking experience replay: a bag of tricks for continual learning. 2020 25th International Conference on Pattern Recognition (ICPR) , 2180–2187.
  4. Retentive or forgetful? diving into the knowledge memorizing mechanism of language models, in: LREC-COLING.
  5. End-to-end incremental learning, in: European Conference on Computer Vision.
  6. Hybrid incremental learning of new data and new classes for hand-held object recognition. Journal of Visual Communication and Image Representation 58, 138–148.
  7. A downsampled variant of imagenet as an alternative to the cifar datasets. ArXiv abs/1707.08819.
  8. Continual prototype evolution: Learning online from non-stationary data streams, in: 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 8230–8239.
  9. An image is worth 16x16 words: Transformers for image recognition at scale, in: ICLR.
  10. Podnet: Pooled outputs distillation for small-tasks incremental learning, in: European Conference on Computer Vision.
  11. Dytox: Transformers for continual learning with dynamic token expansion. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 9275–9285.
  12. Catastrophic interference in connectionist networks: Can it be predicted, can it be prevented?, in: Neural Information Processing Systems.
  13. An empirical investigation of catastrophic forgeting in gradient-based neural networks, in: 2nd International Conference on Learning Representations, ICLR 2014.
  14. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) , 770–778.
  15. Lora: Low-rank adaptation of large language models. ArXiv abs/2106.09685.
  16. Visual prompt tuning. ArXiv abs/2203.12119.
  17. Generating instance-level prompts for rehearsal-free continual learning. 2023 IEEE/CVF International Conference on Computer Vision (ICCV) , 11813–11823.
  18. Learning multiple layers of features from tiny images.
  19. Imagenet classification with deep convolutional neural networks. Communications of the ACM 60, 84 – 90.
  20. Balancing the stability-plasticity dilemma with online stability tuning for continual learning. 2022 International Joint Conference on Neural Networks (IJCNN) , 1–8.
  21. Learning without forgetting. IEEE Transactions on Pattern Analysis and Machine Intelligence 40, 2935–2947.
  22. Cvpr 2020 continual learning in computer vision competition: Approaches, results, current challenges and future directions. Artif. Intell. 303, 103635.
  23. Gradient episodic memory for continual learning, in: Proceedings of the 31st International Conference on Neural Information Processing Systems, p. 6470–6479.
  24. SGDR: Stochastic gradient descent with warm restarts, in: ICLR.
  25. Catastrophic interference in connectionist networks: The sequential learning problem. Psychology of Learning and Motivation 24, 109–165.
  26. The stability-plasticity dilemma: investigating the continuum from catastrophic forgetting to age-limited learning effects. Frontiers in Psychology 4.
  27. Learning transferable visual models from natural language supervision, in: International Conference on Machine Learning.
  28. icarl: Incremental classifier and representation learning, in: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5533–5542.
  29. Boosting separated softmax with discrimination for class incremental learning. Journal of Visual Communication and Image Representation 95, 103899.
  30. Online class-incremental continual learning with adversarial shapley value, in: AAAI Conference on Artificial Intelligence.
  31. Incremental learning of object detectors without catastrophic forgetting, in: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 3420–3429.
  32. Coda-prompt: Continual decomposed attention-based prompting for rehearsal-free continual learning. 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 11909–11919.
  33. Attention is all you need, in: Neural Information Processing Systems.
  34. Three scenarios for continual learning. ArXiv abs/1904.07734.
  35. Three types of incremental learning. Nature Machine Intelligence 4, 1185 – 1197.
  36. Beef: Bi-compatible class-incremental learning via energy-based expansion and fusion, in: International Conference on Learning Representations.
  37. S-prompts learning with pre-trained transformers: An occam’s razor for domain incremental learning, in: NeurIPS.
  38. Isolation and impartial aggregation: A paradigm of incremental learning without interference, in: AAAI Conference on Artificial Intelligence.
  39. DualPrompt: Complementary prompting for rehearsal-free continual learning, in: ECCV.
  40. Learning to prompt for continual learning. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 139–149.
  41. Herding dynamical weights to learn, in: Proceedings of the 26th Annual International Conference on Machine Learning, Association for Computing Machinery, New York, NY, USA. p. 1121–1128.
  42. Cutmix: Regularization strategy to train strong classifiers with localizable features. 2019 IEEE/CVF International Conference on Computer Vision (ICCV) , 6022–6031.
  43. Maintaining discrimination and fairness in class incremental learning. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 13205–13214.
  44. Memory-efficient class-incremental learning for image classification. IEEE Transactions on Neural Networks and Learning Systems 33, 5966–5977.
  45. A model or 603 exemplars: Towards memory-efficient class-incremental learning, in: ICLR.
  46. Revisiting class-incremental learning with pre-trained models: Generalizability and adaptivity are all you need. ArXiv abs/2303.07338.
  47. Learning to prompt for vision-language models. International Journal of Computer Vision 130, 2337 – 2348.
  48. M2kd: Multi-model and multi-level knowledge distillation for incremental learning. ArXiv abs/1904.01769.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets