Regroup Median Loss for Combating Label Noise (2312.06273v1)
Abstract: The deep model training procedure requires large-scale datasets of annotated data. Due to the difficulty of annotating a large number of samples, label noise caused by incorrect annotations is inevitable, resulting in low model performance and poor model generalization. To combat label noise, current methods usually select clean samples based on the small-loss criterion and use these samples for training. Due to some noisy samples similar to clean ones, these small-loss criterion-based methods are still affected by label noise. To address this issue, in this work, we propose Regroup Median Loss (RML) to reduce the probability of selecting noisy samples and correct losses of noisy samples. RML randomly selects samples with the same label as the training samples based on a new loss processing method. Then, we combine the stable mean loss and the robust median loss through a proposed regrouping strategy to obtain robust loss estimation for noisy samples. To further improve the model performance against label noise, we propose a new sample selection strategy and build a semi-supervised method based on RML. Compared to state-of-the-art methods, for both the traditionally trained and semi-supervised models, RML achieves a significant improvement on synthetic and complex real-world datasets. The source code of the paper has been released.
- Unsupervised Label Noise Modeling and Loss Correction. In Proceedings of ICML, June 9-15, 2019, Long Beach, CAL, USA, 312–321.
- A Closer Look at Memorization in Deep Networks. In Proceedings of ICML - Volume 70, 233–242. JMLR.org.
- Understanding and improving early stopping for learning with noisy labels. In Proceedings of NeuIPS, December 6-14, 2021, virtual, 24392–24403.
- MixMatch: A Holistic Approach to Semi-Supervised Learning. In Proceedings of NeuIPS, December 8-14, 2019, Vancouver, BC, Canada, 5050–5060.
- Catoni, O. 2012. Challenging the empirical mean and empirical variance: a deviation study. In Annales de l’IHP Probabilités et statistiques, 1148–1185.
- Learning with Instance-Dependent Label Noise: A Sample Sieve Approach. In Proceedings of ICLR, May 3-7, 2021, Virtual Event, Austria.
- Improving Generalization of Deep Neural Network Acoustic Models with Length Perturbation and N-best Based Label Smoothing. arXiv preprint arXiv:2203.15176.
- Classification in the presence of label noise: a survey. IEEE transactions on neural networks and learning systems, 25: 845–869.
- Girshick, R. 2015. Fast r-cnn. In Proceedings of the CVPR, June 7-12, 2015, Boston, MA, USA, 1440–1448.
- Towards Understanding Deep Learning from Noisy Labels with Small-Loss Criterion. In Proceedings of IJCAI, August 19-27, 2021, Virtual Event, Montreal, Canada, 2469–2475.
- Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Proceedings of NeuIPS, December 3-8, 2018, Montréal, Canada, 8536–8546.
- Mask r-cnn. In Proceedings of the ICCV, October 22-29, 2017, Venice, Italy, 2961–2969.
- Learning with Neighbor Consistency for Noisy Labels. In Proceedings of CVPR, June 18-24, 2022, New Orleans, LA, USA, 4662–4671.
- Deep Bilevel Learning. In Proceedings of CVPR, September 8-14, 2018, Munich, Germany, 632–648.
- UNICON: Combating Label Noise Through Uniform Selection and Contrastive Learning. In Proceedings of the CVPR, June 18-24, 2022, New Orleans, LA, USA, 9676–9686.
- Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60(6): 84–90.
- Robust machine learning by median-of-means: Theory and practice. The Annals of Statistics, 48(2): 906 – 931.
- Dividemix: Learning with noisy labels as semi-supervised learning. In Proceedings of ICLR, April 26-30, 2020, Addis Ababa, Ethiopia, 824–832.
- Learning to Learn From Noisy Labeled Data. In Proceedings of CVPR, June 16-20, 2019, Long Beach, CA, USA, 5051–5059.
- Selective-Supervised Contrastive Learning with Noisy Labels. In Proceedings of CVPR, June 18-24, 2022, New Orleans, LA, USA, 316–325.
- Webvision database: Visual learning and understanding from web data. arXiv preprint arXiv:1708.02862.
- Provably end-to-end label-noise learning without anchor points. In Proceedings of ICML, July 18-24, 2021, Virtual Event, USA, 6403–6413.
- Early-Learning Regularization Prevents Memorization of Noisy Labels. In Proceedings of NeuIPS, December 6-12, 2020, virtual, USA, 20331–20342.
- When does label smoothing help? In Proceedings of NeuIPS, December 8-14, 2019, Vancouver, BC, Canada, 4696–4705.
- Problem complexity and method efficiency in optimization. Siam Review, 27(2): 264–266.
- Multi-objective interpolation training for robustness to label noise. In Proceedings of the CVPR, June 19-25, 2021, virtual, USA, 6606–6615.
- Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach. In Proceedings of CVPR, July 21-26, 2017, Honolulu, HI, USA, 2233–2241.
- Rasmussen, C. 1999. The infinite Gaussian mixture model. In Proceedings of NeuIPS, November 29 - December 4, 1999, Denver, Colorado, USA, volume 12, 554–560.
- You only look once: Unified, real-time object detection. In Proceedings of ICCV, June 27-30, 2016, Las Vegas, NV, USA, 779–788.
- Learning from Noisy Labels with Deep Neural Networks: A Survey. CoRR, abs/2007.08199.
- When ot meets mom: Robust estimation of Wasserstein distance. In Proceedings of AISTATS, April 13 - 15, 2021, Virtual Conference, USA, 136–144.
- Going deeper with convolutions. In Proceedings of CVPR, June 7-12, 2015, Boston, MA, USA, 1–9.
- Joint Optimization Framework for Learning With Noisy Labels. In Proceedings of CVPR, June 18-22, 2018, Salt Lake City, UT, USA, 5552–5560.
- Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of NeuIPS, December 4-9, 2017, Long Beach, CA, USA, 1195–1204.
- Learning with symmetric label noise: The importance of being unhinged. In Proceedings of NeuIPS, December 7-12, 2015, Montreal, Quebec, Canada, 10–18.
- Sample selection with uncertainty of losses for learning with noisy labels. arXiv preprint arXiv:2106.00445.
- Part-dependent label noise: Towards instance-dependent label noise. In Proceedings of NeuIPS, December 6-12, 2020, Virtual Event, USA, 7597–7610.
- Are anchor points really indispensable in label-noise learning? In Proceedings of NeuIPS, December 8-14, 2019, Vancouver, BC, Canada, 6835–6846.
- Learning from massive noisy labeled data for image classification. In Proceedings of CVPR, June 7-12, 2015, Boston, MA, USA, 2691–2699.
- L_DMI: A Novel Information-theoretic Loss Function for Training Deep Nets Robust to Label Noise. In Proceedings of NeuIPS, December 8-14, 2019, Vancouver, BC, Canada, 6222–6233.
- Jo-SRC: A contrastive approach for combating noisy labels. In Proceedings of CVPR, June 19-25, 2021, virtual, USA, 5192–5201.
- Probabilistic End-To-End Noise Correction for Learning With Noisy Labels. In Proceedings of CVPR, June 16-20, 2019, Long Beach, CA, USA, 7017–7025.
- How does disagreement help generalization against label corruption? In Proceedings of ICML, June 9-15, 2019, Long Beach, California, USA, 7164–7173.
- Understanding deep learning requires rethinking generalization. In Proceedings of ICLR.
- Delving deep into label smoothing. IEEE Transactions on Image Processing, 30: 5984–5996.
- Context encoding for semantic segmentation. In Proceedings of CVPR, June 18-22, 2018, Salt Lake City, UT, USA, 7151–7160.
- Accelerating very deep convolutional networks for classification and detection. IEEE transactions on pattern analysis and machine intelligence, 38(10): 1943–1955.
- Learning noise transition matrix from only noisy labels via total variation regularization. In Proceedings of ICML, July 18-24, 2021, Virtual Event, USA, 12501–12512.
- A second-order approach to learning with instance-dependent label noise. In Proceedings of CVPR, June 19-25, 2021, virtual Event, USA, 10113–10123.
- Crowdsourcing the annotation of rumourous conversations in social media. In Proceedings of the WWW, May 18-22, 2015, Florence, Italy, 347–353.