Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Gradient Reweighting: Towards Imbalanced Class-Incremental Learning (2402.18528v2)

Published 28 Feb 2024 in cs.CV

Abstract: Class-Incremental Learning (CIL) trains a model to continually recognize new classes from non-stationary data while retaining learned knowledge. A major challenge of CIL arises when applying to real-world data characterized by non-uniform distribution, which introduces a dual imbalance problem involving (i) disparities between stored exemplars of old tasks and new class data (inter-phase imbalance), and (ii) severe class imbalances within each individual task (intra-phase imbalance). We show that this dual imbalance issue causes skewed gradient updates with biased weights in FC layers, thus inducing over/under-fitting and catastrophic forgetting in CIL. Our method addresses it by reweighting the gradients towards balanced optimization and unbiased classifier learning. Additionally, we observe imbalanced forgetting where paradoxically the instance-rich classes suffer higher performance degradation during CIL due to a larger amount of training data becoming unavailable in subsequent learning phases. To tackle this, we further introduce a distribution-aware knowledge distillation loss to mitigate forgetting by aligning output logits proportionally with the distribution of lost training data. We validate our method on CIFAR-100, ImageNetSubset, and Food101 across various evaluation protocols and demonstrate consistent improvements compared to existing works, showing great potential to apply CIL in real-world scenarios with enhanced robustness and effectiveness.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (54)
  1. Ss-il: Separated softmax for incremental learning. Proceedings of the IEEE/CVF International conference on computer vision, pages 844–853, 2021.
  2. Online continual learning with maximal interfered retrieval. Advances in Neural Information Processing Systems, 32, 2019a.
  3. Gradient based sample selection for online continual learning. Advances in Neural Information Processing Systems, 32, 2019b.
  4. Decomposed knowledge distillation for class-incremental semantic segmentation. Advances in Neural Information Processing Systems, 35:10380–10392, 2022.
  5. Il2m: Class incremental learning with dual memory. Proceedings of the IEEE International Conference on Computer Vision, pages 583–592, 2019.
  6. Active class incremental learning for imbalanced datasets. European Conference on Computer Vision, pages 146–162, 2020.
  7. Food-101 – mining discriminative components with random forests. Proceedings of the European Conference on Computer Vision, 2014.
  8. A systematic study of the class imbalance problem in convolutional neural networks. Neural networks, 106:249–259, 2018.
  9. Learning imbalanced datasets with label-distribution-aware margin loss. Advances in neural information processing systems, 32, 2019.
  10. End-to-end incremental learning. Proceedings of the European Conference on Computer Vision, 2018.
  11. Dynamic residual classifier for class incremental learning. Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 18743–18752, 2023.
  12. Online continual learning from imbalanced data. International Conference on Machine Learning, pages 1952–1961, 2020.
  13. Class-balanced loss based on effective number of samples. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9268–9277, 2019.
  14. Podnet: Pooled outputs distillation for small-tasks incremental learning. Proceedings of the European Conference on Computer Vision, pages 86–102, 2020.
  15. Dealing with cross-task class discrimination in online continual learning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11878–11887, 2023.
  16. Online continual learning for embedded devices. Conference on Lifelong Learning Agents, 2022.
  17. Online continual learning via candidates voting. Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pages 3154–3163, 2022.
  18. Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 770–778, 2016.
  19. Distilling the knowledge in a neural network. Proceedings of the NIPS Deep Learning and Representation Learning Workshop, 2015.
  20. Learning a unified classifier incrementally via rebalancing. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 831–839, 2019.
  21. Multi-level logit distillation. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 24276–24285, 2023.
  22. Imbalanced continual learning with partitioning reservoir sampling. European Conference on Computer Vision, pages 411–428, 2020.
  23. Learning multiple layers of features from tiny images. Technical Report, 2009.
  24. Learning without forgetting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(12):2935–2947, 2017.
  25. Focal loss for dense object detection. Proceedings of the IEEE international conference on computer vision, pages 2980–2988, 2017.
  26. Long-tailed class incremental learning. European Conference on Computer Vision, pages 495–512, 2022.
  27. Mnemonics training: Multi-class incremental learning without forgetting. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 12245–12254, 2020.
  28. Adaptive aggregation networks for class-incremental learning. Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, pages 2544–2553, 2021.
  29. Gradient episodic memory for continual learning. Advances in neural information processing systems, pages 6467–6476, 2017.
  30. Continuous learning in single-incremental-task scenarios. Neural Networks, 116:56–73, 2019.
  31. Class-incremental learning: survey and performance evaluation on image classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(5):5513–5533, 2022.
  32. Catastrophic interference in connectionist networks: The sequential learning problem. pages 109–165. Elsevier, 1989.
  33. Long-tail learning via logit adjustment. International Conference on Learning Representations, 2021.
  34. Influence-balanced loss for imbalanced visual classification. Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 735–744, 2021.
  35. The majority can help the minority: Context-rich minority oversampling for long-tailed classification. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6887–6896, 2022.
  36. Gdumb: A simple approach that questions our progress in continual learning. Proceedings of the European Conference on Computer Vision, pages 524–540, 2020.
  37. iCaRL: Incremental classifier and representation learning. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
  38. Balanced meta-softmax for long-tailed visual recognition. Advances in neural information processing systems, 33:4175–4186, 2020.
  39. ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision, 115(3):211–252, 2015.
  40. On learning the geodesic path for incremental learning. Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, pages 1591–1600, 2021.
  41. Equalization loss for long-tailed object recognition. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 11662–11671, 2020.
  42. Equalization loss v2: A new gradient balance approach for long-tailed object detection. pages 1685–1694, 2021.
  43. Experimental perspectives on learning from imbalanced data. Proceedings of the 24th international conference on Machine learning, pages 935–942, 2007.
  44. Foster: Feature boosting and compression for class-incremental learning. Proceedings of the European Conference on Computer Vision, pages 398–414, 2022.
  45. Max Welling. Herding dynamical weights to learn. Proceedings of the International Conference on Machine Learning, pages 1121–1128, 2009.
  46. Large scale incremental learning. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019.
  47. Der: Dynamically expandable representation for class incremental learning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3014–3023, 2021.
  48. A survey on long-tailed visual recognition. International Journal of Computer Vision, 130(7):1837–1872, 2022.
  49. Lifelong learning with dynamically expandable networks. International Conference on Learning Representations, 2018.
  50. Continual learning through synaptic intelligence. International Conference on Machine Learning, pages 3987–3995, 2017.
  51. Balanced knowledge distillation for long-tailed learning. Neurocomputing, 527:36–46, 2023a.
  52. Deep long-tailed learning: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023b.
  53. Maintaining discrimination and fairness in class incremental learning. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 13208–13217, 2020.
  54. Decoupled knowledge distillation. Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition, pages 11953–11962, 2022.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Jiangpeng He (41 papers)
  2. Fengqing Zhu (76 papers)
Citations (6)