Expandable Subspace Ensemble for Pre-Trained Model-Based Class-Incremental Learning (2403.12030v1)
Abstract: Class-Incremental Learning (CIL) requires a learning system to continually learn new classes without forgetting. Despite the strong performance of Pre-Trained Models (PTMs) in CIL, a critical issue persists: learning new classes often results in the overwriting of old ones. Excessive modification of the network causes forgetting, while minimal adjustments lead to an inadequate fit for new classes. As a result, it is desired to figure out a way of efficient model updating without harming former knowledge. In this paper, we propose ExpAndable Subspace Ensemble (EASE) for PTM-based CIL. To enable model updating without conflict, we train a distinct lightweight adapter module for each new task, aiming to create task-specific subspaces. These adapters span a high-dimensional feature space, enabling joint decision-making across multiple subspaces. As data evolves, the expanding subspaces render the old class classifiers incompatible with new-stage spaces. Correspondingly, we design a semantic-guided prototype complement strategy that synthesizes old classes' new features without using any old class instance. Extensive experiments on seven benchmark datasets verify EASE's state-of-the-art performance. Code is available at: https://github.com/sun-hailong/CVPR24-Ease
- Memory aware synapses: Learning what (not) to forget. In ECCV, pages 139–154, 2018.
- Task-free continual learning. In CVPR, pages 11254–11263, 2019a.
- Gradient based sample selection for online continual learning. In NeurIPS, pages 11816–11825, 2019b.
- Objectnet: A large-scale bias-controlled dataset for pushing the limits of object recognition models. NeurIPS, 32, 2019.
- Il2m: Class incremental learning with dual memory. In ICCV, pages 583–592, 2019.
- Efficient lifelong learning with a-gem. In ICLR, 2018.
- Large-margin contrastive learning with distance polarization regularizer. In ICML, pages 1673–1683, 2021.
- Adaptformer: Adapting vision transformers for scalable visual recognition. In NeurIPS, 2022a.
- Learning contrastive embedding in low-dimensional space. NeurIPS, 35:6345–6357, 2022b.
- Dynamic residual classifier for class incremental learning. In ICCV, pages 18743–18752, 2023.
- Imagenet: A large-scale hierarchical image database. In CVPR, pages 248–255, 2009.
- Learning without memorizing. In CVPR, pages 5138–5146, 2019.
- Federated class-incremental learning. In CVPR, pages 10164–10173, 2022.
- Federated incremental semantic segmentation. In CVPR, pages 3934–3943, 2023.
- An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2020.
- Podnet: Pooled outputs distillation for small-tasks incremental learning. In ECCV, pages 86–102, 2020.
- Dytox: Transformers for continual learning with dynamic token expansion. In CVPR, pages 9285–9295, 2022.
- Robert M French. Catastrophic forgetting in connectionist networks. Trends in cognitive sciences, 3(4):128–135, 1999.
- Modeling time perception in rats: Evidence for catastrophic interference in animal learning. In Proceedings of the 21st Annual Conference of the Cognitive Science Conference, pages 173–178. Citeseer, 1999.
- R-DFCIL: relation-guided representation learning for data-free class incremental learning. In ECCV, pages 423–439, 2022.
- A unified continual learning framework with general parameter-efficient tuning. In ICCV, pages 11483–11493, 2023.
- Fecam: Exploiting the heterogeneity of class distributions in exemplar-free continual learning. NeurIPS, 36, 2023.
- Stephen T Grossberg. Studies of mind and brain: Neural principles of learning, perception, development, cognition, and motor control. Springer Science & Business Media, 2012.
- Pre-trained models: Past, present and future. AI Open, 2:225–250, 2021.
- The many faces of robustness: A critical analysis of out-of-distribution generalization. In ICCV, pages 8340–8349, 2021a.
- Natural adversarial examples. In CVPR, pages 15262–15271, 2021b.
- Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.
- Parameter-efficient transfer learning for nlp. In ICML, pages 2790–2799, 2019.
- Dense network expansion for class incremental learning. In CVPR, pages 11858–11867, 2023.
- Resolving task confusion in dynamic expansion architectures for class incremental learning. In AAAI, pages 908–916, 2023.
- Visual prompt tuning. In ECCV, pages 709–727, 2022.
- Generating instance-level prompts for rehearsal-free continual learning. In ICCV, pages 11847–11857, 2023.
- Leonid V Kantorovich. Mathematical methods of organizing and planning production. Management science, 6(4):366–422, 1960.
- Overcoming catastrophic forgetting in neural networks. PNAS, 114(13):3521–3526, 2017.
- Learning multiple layers of features from tiny images. Technical report, 2009.
- Learning without forgetting. TPAMI, 40(12):2935–2947, 2017.
- Mnemonics training: Multi-class incremental learning without forgetting. In CVPR, pages 12245–12254, 2020.
- Rmm: Reinforced memory management for class-incremental learning. NeurIPS, 34:3478–3490, 2021.
- Premonition: Using generative models to preempt future data changes in continual learning. arXiv preprint arXiv:2403.07356, 2024a.
- Ranpac: Random projections and pre-trained models for continual learning. NeurIPS, 36, 2024b.
- Rf-badge: Vital sign-based authentication via rfid tag array on badges. IEEE Transactions on Mobile Computing, 22(02):1170–1184, 2023.
- Pytorch: An imperative style, high-performance deep learning library. In NeurIPS, pages 8026–8037, 2019.
- Continual normalization: Rethinking batch normalization for online continual learning. In ICLR, 2022.
- Audio-visual class-incremental learning. In ICCV, pages 7799–7811, 2023.
- Roger Ratcliff. Connectionist models of recognition memory: constraints imposed by learning and forgetting functions. Psychological review, 97(2):285, 1990.
- icarl: Incremental classifier and representation learning. In CVPR, pages 2001–2010, 2017.
- Mimicking the oracle: An initial phase decorrelation approach for class incremental learning. In CVPR, pages 16722–16731, 2022.
- On learning the geodesic path for incremental learning. In CVPR, pages 1591–1600, 2021.
- Coda-prompt: Continual decomposed attention-based prompting for rehearsal-free continual learning. In CVPR, pages 11909–11919, 2023.
- Prototypical networks for few-shot learning. In NIPS, pages 4080–4090, 2017.
- Pilot: A pre-trained model-based continual learning toolbox. arXiv preprint arXiv:2309.07117, 2023.
- Topology-preserving class-incremental learning. In ECCV, pages 254–270, 2020.
- Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. JMLR, 9(11), 2008.
- Pivot: Prompting for video continual learning. In CVPR, pages 24214–24223, 2023.
- The Caltech-UCSD Birds-200-2011 Dataset. Technical Report CNS-TR-2011-001, California Institute of Technology, 2011.
- Foster: Feature boosting and compression for class-incremental learning. In ECCV, pages 398–414, 2022a.
- BEEF: Bi-compatible class-incremental learning via energy-based expansion and fusion. In ICLR, 2023a.
- Hierarchical decomposition of prompt-based continual learning: Rethinking obscured sub-optimality. NeurIPS, 36, 2023b.
- Few-shot class-incremental learning via training-free prototype calibration. NeurIPS, 36, 2023c.
- S-prompts learning with pre-trained transformers: An occam’s razor for domain incremental learning. NeurIPS, 35:5682–5695, 2022b.
- Dualprompt: Complementary prompting for rehearsal-free continual learning. In ECCV, pages 631–648, 2022c.
- Learning to prompt for continual learning. In CVPR, pages 139–149, 2022d.
- Large scale incremental learning. In CVPR, pages 374–382, 2019.
- Der: Dynamically expandable representation for class incremental learning. In CVPR, pages 3014–3023, 2021.
- Rectify heterogeneous models with semantic mapping. In ICML, pages 5630–5639, 2018.
- Learning multiple local metrics: Global consideration helps. IEEE transactions on pattern analysis and machine intelligence, 42(7):1698–1712, 2019.
- Semantic drift compensation for class-incremental learning. In CVPR, pages 6982–6991, 2020.
- Continual learning through synaptic intelligence. In ICML, pages 3987–3995, 2017.
- A large-scale study of representation learning with the visual task adaptation benchmark. arXiv preprint arXiv:1910.04867, 2019.
- Slca: Slow learner with classifier alignment for continual learning on a pre-trained model. In ICCV, pages 19148–19158, 2023.
- Class-incremental learning via deep model consolidation. In WACV, pages 1131–1140, 2020.
- Benchmarking omni-vision representation through the lens of visual realms. In ECCV, pages 594–611, 2022.
- Maintaining discrimination and fairness in class incremental learning. In CVPR, pages 13208–13217, 2020.
- Mgsvf: Multi-grained slow versus fast framework for few-shot class-incremental learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 46(3):1576–1588, 2021a.
- Memory-efficient class-incremental learning for image classification. IEEE Transactions on Neural Networks and Learning Systems, 33(10):5966–5977, 2021b.
- Pycil: a python toolbox for class-incremental learning. SCIENCE CHINA Information Sciences, 66(9):197101–, 2023a.
- A model or 603 exemplars: Towards memory-efficient class-incremental learning. In ICLR, 2023b.
- Revisiting class-incremental learning with pre-trained models: Generalizability and adaptivity are all you need. arXiv preprint arXiv:2303.07338, 2023c.
- Continual learning with pre-trained models: A survey. arXiv preprint arXiv:2401.16386, 2024.
- Prototype augmentation and self-supervision for incremental learning. In CVPR, pages 5871–5880, 2021.
- Acil: Analytic class-incremental learning with absolute memorization and privacy protection. NeurIPS, 35:11602–11614, 2022.
- Gkeal: Gaussian kernel embedded analytic learning for few-shot class incremental task. In CVPR, pages 7746–7755, 2023.