Neural Collapse for Cross-entropy Class-Imbalanced Learning with Unconstrained ReLU Feature Model (2401.02058v2)
Abstract: The current paradigm of training deep neural networks for classification tasks includes minimizing the empirical risk that pushes the training loss value towards zero, even after the training error has been vanished. In this terminal phase of training, it has been observed that the last-layer features collapse to their class-means and these class-means converge to the vertices of a simplex Equiangular Tight Frame (ETF). This phenomenon is termed as Neural Collapse (NC). To theoretically understand this phenomenon, recent works employ a simplified unconstrained feature model to prove that NC emerges at the global solutions of the training problem. However, when the training dataset is class-imbalanced, some NC properties will no longer be true. For example, the class-means geometry will skew away from the simplex ETF when the loss converges. In this paper, we generalize NC to imbalanced regime for cross-entropy loss under the unconstrained ReLU feature model. We prove that, while the within-class features collapse property still holds in this setting, the class-means will converge to a structure consisting of orthogonal vectors with different lengths. Furthermore, we find that the classifier weights are aligned to the scaled and centered class-means with scaling factors depend on the number of training samples of each class, which generalizes NC in the class-balanced setting. We empirically prove our results through experiments on practical architectures and dataset.
- On the implicit geometry of cross-entropy parameterizations for label-imbalanced data. In International Conference on Artificial Intelligence and Statistics, pages 10815–10838. PMLR, 2023.
- Learning imbalanced datasets with label-distribution-aware margin loss, 2019.
- Neural collapse in deep linear network: From balanced to imbalanced data. arXiv preprint arXiv:2301.00437, 2023.
- Exploring deep neural networks via layer-peeled model: Minority collapse in imbalanced training. Proceedings of the National Academy of Sciences, 118(43), oct 2021.
- Dissecting supervised contrastive learning, 2023.
- Neural collapse under mse loss: Proximity to and dynamics on the central path, 2021.
- Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pages 770–778. IEEE Computer Society, 2016.
- W. Hong and S. Ling. Neural collapse for unconstrained feature model under cross-entropy loss with imbalanced data. arXiv preprint arXiv:2309.09725, 2023.
- K. Hornik. Approximation capabilities of multilayer feedforward networks. Neural Networks, 4(2):251–257, 1991.
- Multilayer feedforward networks are universal approximators. Neural Networks, 2(5):359–366, 1989.
- Learning deep representation for imbalanced classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5375–5384, 2016.
- Exploring balanced feature spaces for representation learning. In International Conference on Learning Representations, 2020.
- Decoupling representation and classifier for long-tailed recognition, 2019.
- B. Kim and J. Kim. Adjusting decision boundary for class imbalanced learning. IEEE Access, 8:81674–81685, 2020.
- Supervised-contrastive loss learns orthogonal frames and batching matters. arXiv preprint arXiv:2306.07960, 2023.
- Learning multiple layers of features from tiny images, 2009.
- Inducing neural collapse in deep long-tailed learning. In International Conference on Artificial Intelligence and Statistics, pages 11534–11544. PMLR, 2023.
- J. Lu and S. Steinerberger. Neural collapse with cross-entropy loss, 2020.
- Memorization-dilation: Modeling neural collapse under noise. In The Eleventh International Conference on Learning Representations, 2022.
- Prevalence of neural collapse during the terminal phase of deep learning training. CoRR, abs/2008.08186, 2020.
- Deep neural collapse is provably optimal for the deep unconstrained features model. arXiv preprint arXiv:2305.13165, 2023.
- Imbalance trouble: Revisiting neural-collapse geometry, 2022.
- T. Tirer and J. Bruna. Extended unconstrained features model for exploring deep neural collapse, 2022.
- Perturbation analysis of neural collapse. In International Conference on Machine Learning, pages 34301–34329. PMLR, 2023.
- Inducing neural collapse in imbalanced learning: Do we really need a learnable classifier at the end of deep neural network?, 2022.
- D. Yarotsky. Universal approximations of invariant maps by neural networks, 2018.
- Identifying and compensating for feature deviation in imbalanced deep learning. arXiv preprint arXiv:2001.01385, 2020.
- D.-X. Zhou. Universality of deep convolutional neural networks, 2018.
- On the optimization landscape of neural collapse under mse loss: Global optimality with unconstrained features, 2022.
- Are all losses created equal: A neural collapse perspective, 2022.
- A geometric analysis of neural collapse with unconstrained features. CoRR, abs/2105.02375, 2021.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.