Gradient Reweighting: Towards Imbalanced Class-Incremental Learning (2402.18528v2)
Abstract: Class-Incremental Learning (CIL) trains a model to continually recognize new classes from non-stationary data while retaining learned knowledge. A major challenge of CIL arises when applying to real-world data characterized by non-uniform distribution, which introduces a dual imbalance problem involving (i) disparities between stored exemplars of old tasks and new class data (inter-phase imbalance), and (ii) severe class imbalances within each individual task (intra-phase imbalance). We show that this dual imbalance issue causes skewed gradient updates with biased weights in FC layers, thus inducing over/under-fitting and catastrophic forgetting in CIL. Our method addresses it by reweighting the gradients towards balanced optimization and unbiased classifier learning. Additionally, we observe imbalanced forgetting where paradoxically the instance-rich classes suffer higher performance degradation during CIL due to a larger amount of training data becoming unavailable in subsequent learning phases. To tackle this, we further introduce a distribution-aware knowledge distillation loss to mitigate forgetting by aligning output logits proportionally with the distribution of lost training data. We validate our method on CIFAR-100, ImageNetSubset, and Food101 across various evaluation protocols and demonstrate consistent improvements compared to existing works, showing great potential to apply CIL in real-world scenarios with enhanced robustness and effectiveness.
- Ss-il: Separated softmax for incremental learning. Proceedings of the IEEE/CVF International conference on computer vision, pages 844–853, 2021.
- Online continual learning with maximal interfered retrieval. Advances in Neural Information Processing Systems, 32, 2019a.
- Gradient based sample selection for online continual learning. Advances in Neural Information Processing Systems, 32, 2019b.
- Decomposed knowledge distillation for class-incremental semantic segmentation. Advances in Neural Information Processing Systems, 35:10380–10392, 2022.
- Il2m: Class incremental learning with dual memory. Proceedings of the IEEE International Conference on Computer Vision, pages 583–592, 2019.
- Active class incremental learning for imbalanced datasets. European Conference on Computer Vision, pages 146–162, 2020.
- Food-101 – mining discriminative components with random forests. Proceedings of the European Conference on Computer Vision, 2014.
- A systematic study of the class imbalance problem in convolutional neural networks. Neural networks, 106:249–259, 2018.
- Learning imbalanced datasets with label-distribution-aware margin loss. Advances in neural information processing systems, 32, 2019.
- End-to-end incremental learning. Proceedings of the European Conference on Computer Vision, 2018.
- Dynamic residual classifier for class incremental learning. Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 18743–18752, 2023.
- Online continual learning from imbalanced data. International Conference on Machine Learning, pages 1952–1961, 2020.
- Class-balanced loss based on effective number of samples. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9268–9277, 2019.
- Podnet: Pooled outputs distillation for small-tasks incremental learning. Proceedings of the European Conference on Computer Vision, pages 86–102, 2020.
- Dealing with cross-task class discrimination in online continual learning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11878–11887, 2023.
- Online continual learning for embedded devices. Conference on Lifelong Learning Agents, 2022.
- Online continual learning via candidates voting. Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pages 3154–3163, 2022.
- Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 770–778, 2016.
- Distilling the knowledge in a neural network. Proceedings of the NIPS Deep Learning and Representation Learning Workshop, 2015.
- Learning a unified classifier incrementally via rebalancing. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 831–839, 2019.
- Multi-level logit distillation. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 24276–24285, 2023.
- Imbalanced continual learning with partitioning reservoir sampling. European Conference on Computer Vision, pages 411–428, 2020.
- Learning multiple layers of features from tiny images. Technical Report, 2009.
- Learning without forgetting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(12):2935–2947, 2017.
- Focal loss for dense object detection. Proceedings of the IEEE international conference on computer vision, pages 2980–2988, 2017.
- Long-tailed class incremental learning. European Conference on Computer Vision, pages 495–512, 2022.
- Mnemonics training: Multi-class incremental learning without forgetting. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 12245–12254, 2020.
- Adaptive aggregation networks for class-incremental learning. Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, pages 2544–2553, 2021.
- Gradient episodic memory for continual learning. Advances in neural information processing systems, pages 6467–6476, 2017.
- Continuous learning in single-incremental-task scenarios. Neural Networks, 116:56–73, 2019.
- Class-incremental learning: survey and performance evaluation on image classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(5):5513–5533, 2022.
- Catastrophic interference in connectionist networks: The sequential learning problem. pages 109–165. Elsevier, 1989.
- Long-tail learning via logit adjustment. International Conference on Learning Representations, 2021.
- Influence-balanced loss for imbalanced visual classification. Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 735–744, 2021.
- The majority can help the minority: Context-rich minority oversampling for long-tailed classification. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6887–6896, 2022.
- Gdumb: A simple approach that questions our progress in continual learning. Proceedings of the European Conference on Computer Vision, pages 524–540, 2020.
- iCaRL: Incremental classifier and representation learning. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
- Balanced meta-softmax for long-tailed visual recognition. Advances in neural information processing systems, 33:4175–4186, 2020.
- ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision, 115(3):211–252, 2015.
- On learning the geodesic path for incremental learning. Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, pages 1591–1600, 2021.
- Equalization loss for long-tailed object recognition. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 11662–11671, 2020.
- Equalization loss v2: A new gradient balance approach for long-tailed object detection. pages 1685–1694, 2021.
- Experimental perspectives on learning from imbalanced data. Proceedings of the 24th international conference on Machine learning, pages 935–942, 2007.
- Foster: Feature boosting and compression for class-incremental learning. Proceedings of the European Conference on Computer Vision, pages 398–414, 2022.
- Max Welling. Herding dynamical weights to learn. Proceedings of the International Conference on Machine Learning, pages 1121–1128, 2009.
- Large scale incremental learning. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019.
- Der: Dynamically expandable representation for class incremental learning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3014–3023, 2021.
- A survey on long-tailed visual recognition. International Journal of Computer Vision, 130(7):1837–1872, 2022.
- Lifelong learning with dynamically expandable networks. International Conference on Learning Representations, 2018.
- Continual learning through synaptic intelligence. International Conference on Machine Learning, pages 3987–3995, 2017.
- Balanced knowledge distillation for long-tailed learning. Neurocomputing, 527:36–46, 2023a.
- Deep long-tailed learning: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023b.
- Maintaining discrimination and fairness in class incremental learning. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 13208–13217, 2020.
- Decoupled knowledge distillation. Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition, pages 11953–11962, 2022.
- Jiangpeng He (41 papers)
- Fengqing Zhu (76 papers)