Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 167 tok/s
Gemini 2.5 Pro 53 tok/s Pro
GPT-5 Medium 31 tok/s Pro
GPT-5 High 31 tok/s Pro
GPT-4o 106 tok/s Pro
Kimi K2 187 tok/s Pro
GPT OSS 120B 443 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Neural Collapse for Cross-entropy Class-Imbalanced Learning with Unconstrained ReLU Feature Model (2401.02058v2)

Published 4 Jan 2024 in cs.LG and stat.ML

Abstract: The current paradigm of training deep neural networks for classification tasks includes minimizing the empirical risk that pushes the training loss value towards zero, even after the training error has been vanished. In this terminal phase of training, it has been observed that the last-layer features collapse to their class-means and these class-means converge to the vertices of a simplex Equiangular Tight Frame (ETF). This phenomenon is termed as Neural Collapse (NC). To theoretically understand this phenomenon, recent works employ a simplified unconstrained feature model to prove that NC emerges at the global solutions of the training problem. However, when the training dataset is class-imbalanced, some NC properties will no longer be true. For example, the class-means geometry will skew away from the simplex ETF when the loss converges. In this paper, we generalize NC to imbalanced regime for cross-entropy loss under the unconstrained ReLU feature model. We prove that, while the within-class features collapse property still holds in this setting, the class-means will converge to a structure consisting of orthogonal vectors with different lengths. Furthermore, we find that the classifier weights are aligned to the scaled and centered class-means with scaling factors depend on the number of training samples of each class, which generalizes NC in the class-balanced setting. We empirically prove our results through experiments on practical architectures and dataset.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. On the implicit geometry of cross-entropy parameterizations for label-imbalanced data. In International Conference on Artificial Intelligence and Statistics, pages 10815–10838. PMLR, 2023.
  2. Learning imbalanced datasets with label-distribution-aware margin loss, 2019.
  3. Neural collapse in deep linear network: From balanced to imbalanced data. arXiv preprint arXiv:2301.00437, 2023.
  4. Exploring deep neural networks via layer-peeled model: Minority collapse in imbalanced training. Proceedings of the National Academy of Sciences, 118(43), oct 2021.
  5. Dissecting supervised contrastive learning, 2023.
  6. Neural collapse under mse loss: Proximity to and dynamics on the central path, 2021.
  7. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pages 770–778. IEEE Computer Society, 2016.
  8. W. Hong and S. Ling. Neural collapse for unconstrained feature model under cross-entropy loss with imbalanced data. arXiv preprint arXiv:2309.09725, 2023.
  9. K. Hornik. Approximation capabilities of multilayer feedforward networks. Neural Networks, 4(2):251–257, 1991.
  10. Multilayer feedforward networks are universal approximators. Neural Networks, 2(5):359–366, 1989.
  11. Learning deep representation for imbalanced classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5375–5384, 2016.
  12. Exploring balanced feature spaces for representation learning. In International Conference on Learning Representations, 2020.
  13. Decoupling representation and classifier for long-tailed recognition, 2019.
  14. B. Kim and J. Kim. Adjusting decision boundary for class imbalanced learning. IEEE Access, 8:81674–81685, 2020.
  15. Supervised-contrastive loss learns orthogonal frames and batching matters. arXiv preprint arXiv:2306.07960, 2023.
  16. Learning multiple layers of features from tiny images, 2009.
  17. Inducing neural collapse in deep long-tailed learning. In International Conference on Artificial Intelligence and Statistics, pages 11534–11544. PMLR, 2023.
  18. J. Lu and S. Steinerberger. Neural collapse with cross-entropy loss, 2020.
  19. Memorization-dilation: Modeling neural collapse under noise. In The Eleventh International Conference on Learning Representations, 2022.
  20. Prevalence of neural collapse during the terminal phase of deep learning training. CoRR, abs/2008.08186, 2020.
  21. Deep neural collapse is provably optimal for the deep unconstrained features model. arXiv preprint arXiv:2305.13165, 2023.
  22. Imbalance trouble: Revisiting neural-collapse geometry, 2022.
  23. T. Tirer and J. Bruna. Extended unconstrained features model for exploring deep neural collapse, 2022.
  24. Perturbation analysis of neural collapse. In International Conference on Machine Learning, pages 34301–34329. PMLR, 2023.
  25. Inducing neural collapse in imbalanced learning: Do we really need a learnable classifier at the end of deep neural network?, 2022.
  26. D. Yarotsky. Universal approximations of invariant maps by neural networks, 2018.
  27. Identifying and compensating for feature deviation in imbalanced deep learning. arXiv preprint arXiv:2001.01385, 2020.
  28. D.-X. Zhou. Universality of deep convolutional neural networks, 2018.
  29. On the optimization landscape of neural collapse under mse loss: Global optimality with unconstrained features, 2022.
  30. Are all losses created equal: A neural collapse perspective, 2022.
  31. A geometric analysis of neural collapse with unconstrained features. CoRR, abs/2105.02375, 2021.
Citations (9)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 15 likes.

Upgrade to Pro to view all of the tweets about this paper: