Class incremental learning with probability dampening and cascaded gated classifier
Abstract: Humans are capable of acquiring new knowledge and transferring learned knowledge into different domains, incurring a small forgetting. The same ability, called Continual Learning, is challenging to achieve when operating with neural networks due to the forgetting affecting past learned tasks when learning new ones. This forgetting can be mitigated by replaying stored samples from past tasks, but a large memory size may be needed for long sequences of tasks; moreover, this could lead to overfitting on saved samples. In this paper, we propose a novel regularisation approach and a novel incremental classifier called, respectively, Margin Dampening and Cascaded Scaling Classifier. The first combines a soft constraint and a knowledge distillation approach to preserve past learned knowledge while allowing the model to learn new patterns effectively. The latter is a gated incremental classifier, helping the model modify past predictions without directly interfering with them. This is achieved by modifying the output of the model with auxiliary scaling functions. We empirically show that our approach performs well on multiple benchmarks against well-established baselines, and we also study each component of our proposal and how the combinations of such components affect the final results.
- Ss-il: Separated softmax for incremental learning. In Proceedings of the IEEE/CVF International conference on computer vision, pp. 844–853, 2021.
- Memory Aware Synapses: Learning What (not) to Forget, pp. 144–161. Springer International Publishing, 2018.
- Rainbow memory: Continual learning with a memory of diverse samples. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8218–8227, 2021.
- On the effectiveness of lipschitz-driven rehearsal in continual learning. Advances in Neural Information Processing Systems, 35:31886–31901, 2022.
- Dark experience for general continual learning: a strong, simple baseline. Advances in neural information processing systems, 33:15920–15930, 2020.
- Rethinking experience replay: a bag of tricks for continual learning. In 2020 25th International Conference on Pattern Recognition (ICPR), pp. 2180–2187. IEEE, 2021.
- New insights on reducing abrupt representation change in online continual learning. In International Conference on Learning Representations, 2022.
- On tiny episodic memories in continual learning. arXiv preprint arXiv:1902.10486, 2019.
- Using hindsight to anchor past knowledge in continual learning. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pp. 6993–7001, 2021.
- Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255, 2009.
- Don’t forget, there is more than forgetting: new metrics for continual learning. arXiv preprint arXiv:1810.13166, 2018.
- Podnet: Pooled outputs distillation for small-tasks incremental learning. In Computer Vision–ECCV 2020: 16th European Conference, pp. 86–102. Springer, 2020.
- Casper: Latent spectral regularization for continual learning. arXiv preprint arXiv:2301.03345, 2023.
- Ddgr: Continual learning with deep diffusion-based generative replay. In Proceedings of the 40th International Conference on Machine Learning, ICML’23. JMLR.org, 2023.
- Continual learning via neural pruning. In Real Neurons & Hidden Units: Future directions at the intersection of neuroscience and artificial intelligence @ NeurIPS 2019, 2019.
- Continually learning self-supervised representations with projected functional regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3867–3877, 2022.
- Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.
- Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences, 114(13):3521–3526, 2017. ISSN 1091-6490.
- Continuous domain adaptation with variational domain-agnostic feature replay. arXiv preprint arXiv:2003.04382, 2020.
- Learning without forgetting. IEEE transactions on pattern analysis and machine intelligence, 40(12):2935–2947, 2017.
- Loss decoupling for task-agnostic continual learning. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
- Avalanche: an end-to-end library for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3600–3610, 2021.
- Gradient episodic memory for continual learning. Advances in neural information processing systems, 30, 2017.
- Packnet: Adding multiple tasks to a single network by iterative pruning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 7765–7773, 2018.
- Unified probabilistic deep continual learning through generative replay and open set recognition. Journal of Imaging, 8(4), 2022.
- Continual lifelong learning with neural networks: A review. Neural Networks, 113:54–71, 2019.
- Class-incremental learning with pre-allocated fixed classifiers. In 2020 25th International Conference on Pattern Recognition (ICPR), pp. 6259–6266. IEEE, 2021.
- Continual learning, fast and slow. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
- Structured ensembles: An approach to reduce the memory footprint of ensemble methods. Neural Networks, 144:407–418, 2021. ISSN 0893-6080.
- Centroids matching: an efficient continual learning approach operating in the embedding space. Transactions on Machine Learning Research, 2022. ISSN 2835-8856.
- Continual learning with invertible generative models. Neural Networks, 164:606–616, 2023.
- Gdumb: A simple approach that questions our progress in continual learning. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part II 16, pp. 524–540. Springer, 2020.
- Progressive neural networks. arXiv preprint arXiv:1606.04671, 2016.
- Brain-inspired replay for continual learning with artificial neural networks. Nature communications, 11(1):4069, 2020.
- Rehearsal revealed: The limits and merits of revisiting samples in continual learning. In 2021 IEEE/CVF International Conference on Computer Vision (ICCV). IEEE, 2021.
- A comprehensive survey of continual learning: Theory, method and application. arXiv preprint arXiv:2302.00487, 2023.
- Sparcl: Sparse continual learning on the edge. Advances in Neural Information Processing Systems, 35:20366–20380, 2022.
- Supermasks in superposition. Advances in Neural Information Processing Systems, 33:15173–15184, 2020.
- Large scale incremental learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 374–382, 2019.
- Continual learning via manifold expansion replay. arXiv preprint arXiv:2310.08038, 2023.
- Continual learning through synaptic intelligence. In International conference on machine learning, pp. 3987–3995. PMLR, 2017.
- Masked autoencoders are efficient class incremental learners. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 19104–19113, 2023.
- Maintaining discrimination and fairness in class incremental learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 13208–13217, 2020.
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.