Learning with Imbalanced Noisy Data by Preventing Bias in Sample Selection (2402.11242v1)
Abstract: Learning with noisy labels has gained increasing attention because the inevitable imperfect labels in real-world scenarios can substantially hurt the deep model performance. Recent studies tend to regard low-loss samples as clean ones and discard high-loss ones to alleviate the negative impact of noisy labels. However, real-world datasets contain not only noisy labels but also class imbalance. The imbalance issue is prone to causing failure in the loss-based sample selection since the under-learning of tail classes also leans to produce high losses. To this end, we propose a simple yet effective method to address noisy labels in imbalanced datasets. Specifically, we propose Class-Balance-based sample Selection (CBS) to prevent the tail class samples from being neglected during training. We propose Confidence-based Sample Augmentation (CSA) for the chosen clean samples to enhance their reliability in the training process. To exploit selected noisy samples, we resort to prediction history to rectify labels of noisy samples. Moreover, we introduce the Average Confidence Margin (ACM) metric to measure the quality of corrected labels by leveraging the model's evolving training dynamics, thereby ensuring that low-quality corrected noisy samples are appropriately masked out. Lastly, consistency regularization is imposed on filtered label-corrected noisy samples to boost model performance. Comprehensive experimental results on synthetic and real-world datasets demonstrate the effectiveness and superiority of our proposed method, especially in imbalanced scenarios. Comprehensive experimental results on synthetic and real-world datasets demonstrate the effectiveness and superiority of our proposed method, especially in imbalanced scenarios.
- X. Zhong, C. Gu, M. Ye, W. Huang, and C. Lin, “Graph complemented latent representation for few-shot image classification,” IEEE Trans. Multimedia, vol. 25, pp. 1979–1990, 2023.
- A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Adv. Neural Inform. Process. Syst., 2012, pp. 1106–1114.
- Z. Shao, Y. Pu, J. Zhou, B. Wen, and Y. Zhang, “Hyper RPCA: joint maximum correntropy criterion and laplacian scale mixture modeling on-the-fly for moving object detection,” IEEE Trans. Multimedia, vol. 25, pp. 112–125, 2023.
- J. Redmon and A. Farhadi, “YOLO9000: better, faster, stronger,” in IEEE Conf. Comput. Vis. Pattern Recog., 2017, pp. 6517–6525.
- F. Boutros, N. Damer, F. Kirchbuchner, and A. Kuijper, “Elasticface: Elastic margin loss for deep face recognition,” in IEEE Conf. Comput. Vis. Pattern Recog., 2022, pp. 1577–1586.
- J. Qian, S. Zhu, C. Zhao, J. Yang, and W. K. Wong, “Otface: Hard samples guided optimal transport loss for deep face representation,” IEEE Trans. Multimedia, vol. 25, pp. 1427–1438, 2023.
- H. Ying, Z. Huang, S. Liu, T. Shao, and K. Zhou, “Embedmask: Embedding coupling for instance segmentation,” in IJCAI, 2021, pp. 1266–1273.
- K. Zhang, C. Yuan, Y. Zhu, Y. Jiang, and L. Luo, “Weakly supervised instance segmentation by exploring entire object regions,” IEEE Trans. Multimedia, vol. 25, pp. 352–363, 2023.
- T. Chen, Y. Yao, and J. Tang, “Multi-granularity denoising and bidirectional alignment for weakly supervised semantic segmentation,” IEEE Trans. Image Process., vol. 32, pp. 2960–2971, 2023.
- T. Chen, Y. Yao, L. Zhang, Q. Wang, G. Xie, and F. Shen, “Saliency guided inter- and intra-class relation constraints for weakly supervised semantic segmentation,” IEEE Trans. Multimedia, vol. 25, pp. 1727–1737, 2023.
- E. L. Malfa, R. Michelmore, A. M. Zbrzezny, N. Paoletti, and M. Kwiatkowska, “On guaranteed optimal robust explanations for NLP models,” in IJCAI, 2021, pp. 2658–2665.
- J. Deng, W. Dong, R. Socher, L. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in IEEE Conf. Comput. Vis. Pattern Recog., 2009, pp. 248–255.
- T. Wu, B. Dai, S. Chen, Y. Qu, and Y. Xie, “Meta segmentation network for ultra-resolution medical images,” in IJCAI, 2020, pp. 544–550.
- P. Welinder, S. Branson, S. J. Belongie, and P. Perona, “The multidimensional wisdom of crowds,” in Adv. Neural Inform. Process. Syst., 2010, pp. 2424–2432.
- R. Fergus, L. Fei-Fei, P. Perona, and A. Zisserman, “Learning object categories from internet image searches,” Proc. IEEE, pp. 1453–1466, 2010.
- C. Zhang, S. Bengio, M. Hardt, B. Recht, and O. Vinyals, “Understanding deep learning requires rethinking generalization,” in Int. Conf. Learn. Represent., 2017.
- B. Han, Q. Yao, X. Yu, G. Niu, M. Xu, W. Hu, I. W. Tsang, and M. Sugiyama, “Co-teaching: Robust training of deep neural networks with extremely noisy labels,” in Adv. Neural Inform. Process. Syst., 2018, pp. 8536–8546.
- Y. Yao, Z. Sun, C. Zhang, F. Shen, Q. Wu, J. Zhang, and Z. Tang, “Jo-src: A contrastive approach for combating noisy labels,” in IEEE Conf. Comput. Vis. Pattern Recog., 2021, pp. 5192–5201.
- H. Wei, L. Feng, X. Chen, and B. An, “Combating noisy labels by agreement: A joint training method with co-regularization,” in IEEE Conf. Comput. Vis. Pattern Recog., 2020, pp. 13 723–13 732.
- X. Yu, B. Han, J. Yao, G. Niu, I. W. Tsang, and M. Sugiyama, “How does disagreement help generalization against label corruption?” in Int. Conf. Mach. Learn., 2019, pp. 7164–7173.
- J. Li, R. Socher, and S. C. Hoi, “Dividemix: Learning with noisy labels as semi-supervised learning,” in Int. Conf. Learn. Represent., 2020.
- Z. Sun, F. Shen, D. Huang, Q. Wang, X. Shu, Y. Yao, and J. Tang, “Pnp: Robust learning from noisy labels by probabilistic noise prediction,” in IEEE Conf. Comput. Vis. Pattern Recog., 2022, pp. 5311–5320.
- Y. Lu and W. He, “SELC: self-ensemble label correction improves learning with noisy labels,” in IJCAI, 2022, pp. 3278–3284.
- Y. Liu, N. Xu, Y. Zhang, and X. Geng, “Label distribution for learning with noisy labels,” in IJCAI, 2020, pp. 2568–2574.
- Y. Bai, E. Yang, B. Han, Y. Yang, J. Li, Y. Mao, G. Niu, and T. Liu, “Understanding and improving early stopping for learning with noisy labels,” in Adv. Neural Inform. Process. Syst., 2021, pp. 24 392–24 403.
- X. Xia, B. Han, N. Wang, J. Deng, J. Li, Y. Mao, and T. Liu, “Extended $t$t: Learning with mixed closed-set and open-set noisy labels,” IEEE Trans. Pattern Anal. Mach. Intell., pp. 3047–3058, 2023.
- D. Cheng, T. Liu, Y. Ning, N. Wang, B. Han, G. Niu, X. Gao, and M. Sugiyama, “Instance-dependent label-noise learning with manifold-regularized transition matrix estimation,” in IEEE Conf. Comput. Vis. Pattern Recog., 2022, pp. 16 609–16 618.
- S. Li, X. Xia, S. Ge, and T. Liu, “Selective-supervised contrastive learning with noisy labels,” in IEEE Conf. Comput. Vis. Pattern Recog., 2022, pp. 316–325.
- E. Yang, D. Yao, T. Liu, and C. Deng, “Mutual quantization for cross-modal search with noisy labels,” in IEEE Conf. Comput. Vis. Pattern Recog., 2022, pp. 7541–7550.
- C. Gong, Y. Ding, B. Han, G. Niu, J. Yang, J. You, D. Tao, and M. Sugiyama, “Class-wise denoising for robust learning under label noise,” IEEE Trans. Pattern Anal. Mach. Intell., pp. 2835–2848, 2023.
- E. Arazo, D. Ortego, P. Albert, N. E. O’Connor, and K. McGuinness, “Unsupervised label noise modeling and loss correction,” in Int. Conf. Mach. Learn., 2019, pp. 312–321.
- K. Yi and J. Wu, “Probabilistic end-to-end noise correction for learning with noisy labels,” in IEEE Conf. Comput. Vis. Pattern Recog., 2019, pp. 7017–7025.
- Z. Sun, Y. Yao, X. Wei, F. Shen, H. Liu, and X.-S. Hua, “Boosting robust learning via leveraging reusable samples in noisy web data,” IEEE Trans. Multimedia, 2022.
- G. Patrini, A. Rozza, A. Krishna Menon, R. Nock, and L. Qu, “Making deep neural networks robust to label noise: A loss correction approach,” in IEEE Conf. Comput. Vis. Pattern Recog., 2017, pp. 1944–1952.
- J. Goldberger and E. Ben-Reuven, “Training deep neural-networks using a noise adaptation layer,” in Int. Conf. Learn. Represent., 2017.
- X. Xia, T. Liu, N. Wang, B. Han, C. Gong, G. Niu, and M. Sugiyama, “Are anchor points really indispensable in label-noise learning?” in Adv. Neural Inform. Process. Syst., H. M. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. B. Fox, and R. Garnett, Eds., 2019, pp. 6835–6846.
- Z. Sun, Y. Yao, X.-S. Wei, Y. Zhang, F. Shen, J. Wu, J. Zhang, and H.-T. Shen, “Webly supervised fine-grained recognition: Benchmark datasets and an approach,” in Int. Conf. Comput. Vis., 2021, pp. 10 602–10 611.
- X. Gui, W. Wang, and Z. Tian, “Towards understanding deep learning from noisy labels with small-loss criterion,” in IJCAI, 2021, pp. 2469–2475.
- Y. Cui, M. Jia, T. Lin, Y. Song, and S. J. Belongie, “Class-balanced loss based on effective number of samples,” in IEEE Conf. Comput. Vis. Pattern Recog., 2019, pp. 9268–9277.
- T. Xiao, T. Xia, Y. Yang, C. Huang, and X. Wang, “Learning from massive noisy labeled data for image classification,” in IEEE Conf. Comput. Vis. Pattern Recog., 2015, pp. 2691–2699.
- Z. Sun, X.-S. Hua, Y. Yao, X.-S. Wei, G. Hu, and J. Zhang, “Crssc: salvage reusable samples from noisy data for robust learning,” in ACM Int. Conf. Multimedia, 2020, pp. 92–101.
- C. Zhang, Y. Yao, X. Xu, J. Shao, J. Song, Z. Li, and Z. Tang, “Extracting useful knowledge from noisy web images via data purification for fine-grained recognition,” in ACM Int. Conf. Multimedia, 2021, pp. 4063–4072.
- H. Liu, C. Zhang, Y. Yao, X.-S. Wei, F. Shen, Z. Tang, and J. Zhang, “Exploiting web images for fine-grained visual recognition by eliminating open-set noise and utilizing hard examples,” IEEE Trans. Multimedia, vol. 24, pp. 546–557, 2021.
- Y. Yao, X. Hua, G. Gao, Z. Sun, Z. Li, and J. Zhang, “Bridging the web data and fine-grained visual recognition via alleviating label noise and domain mismatch,” in ACM Int. Conf. Multimedia, 2020, pp. 1735–1744.
- Z. Zhang and M. R. Sabuncu, “Generalized cross entropy loss for training deep neural networks with noisy labels,” in Adv. Neural Inform. Process. Syst., 2018, pp. 8792–8802.
- X. Zhou, X. Liu, D. Zhai, J. Jiang, and X. Ji, “Asymmetric loss functions for noise-tolerant learning: Theory and applications,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 7, pp. 8094–8109, 2023.
- H. Zhang, M. Cissé, Y. N. Dauphin, and D. Lopez-Paz, “mixup: Beyond empirical risk minimization,” in Int. Conf. Learn. Represent., 2018.
- S. Liu, J. Niles-Weed, N. Razavian, and C. Fernandez-Granda, “Early-learning regularization prevents memorization of noisy labels,” in Adv. Neural Inform. Process. Syst., 2020.
- E. Zheltonozhskii, C. Baskin, A. Mendelson, A. M. Bronstein, and O. Litany, “Contrast to divide: Self-supervised pre-training for learning with noisy labels,” in IEEE Winter Conference on Applications of Computer Vision, 2022, pp. 387–397.
- C. Zhang, G. Lin, Q. Wang, F. Shen, Y. Yao, and Z. Tang, “Guided by meta-set: A data-driven method for fine-grained visual recognition,” IEEE Trans. Multimedia, vol. 25, pp. 4691–4703, 2023.
- D. Tanaka, D. Ikami, T. Yamasaki, and K. Aizawa, “Joint optimization framework for learning with noisy labels,” in IEEE Conf. Comput. Vis. Pattern Recog., 2018, pp. 5552–5560.
- J. Li, G. Li, F. Liu, and Y. Yu, “Neighborhood collective estimation for noisy label identification and correction,” in Eur. Conf. Comput. Vis., 2022, pp. 128–145.
- D. Patel and P. S. Sastry, “Adaptive sample selection for robust learning under label noise,” in IEEE Winter Conference on Applications of Computer Vision, 2023, pp. 3921–3931.
- X. Zhou, X. Liu, C. Wang, D. Zhai, J. Jiang, and X. Ji, “Learning with noisy labels via sparse regularization,” in Int. Conf. Comput. Vis., 2021, pp. 72–81.
- Z. Huang, J. Zhang, and H. Shan, “Twin contrastive learning with noisy labels,” in IEEE Conf. Comput. Vis. Pattern Recog., 2023, pp. 11 661–11 670.
- M. Ren, W. Zeng, B. Yang, and R. Urtasun, “Learning to reweight examples for robust deep learning,” in Int. Conf. Mach. Learn., 2018, pp. 4331–4340.
- J. Shu, Q. Xie, L. Yi, Q. Zhao, S. Zhou, Z. Xu, and D. Meng, “Meta-weight-net: Learning an explicit mapping for sample weighting,” in Adv. Neural Inform. Process. Syst., 2019, pp. 1917–1928.
- S. Jiang, J. Li, Y. Wang, B. Huang, Z. Zhang, and T. Xu, “Delving into sample loss curve to embrace noisy and imbalanced data,” in AAAI, 2022, pp. 7024–7032.
- Y. Huang, B. Bai, S. Zhao, K. Bai, and F. Wang, “Uncertainty-aware learning against label noise on imbalanced datasets,” in AAAI, 2022, pp. 6960–6969.
- X. Xia, T. Liu, B. Han, M. Gong, J. Yu, G. Niu, and M. Sugiyama, “Sample selection with uncertainty of losses for learning with noisy labels,” in Int. Conf. Learn. Represent., 2022.
- X. Xia, B. Han, Y. Zhan, J. Yu, M. Gong, C. Gong, and T. Liu, “Combating noisy labels with sample selection by mining high-discrepancy examples,” in Int. Conf. Comput. Vis., 2023, pp. 1833–1843.
- N. Karim, M. N. Rizve, N. Rahnavard, A. Mian, and M. Shah, “UNICON: combating label noise through uniform selection and contrastive learning,” in IEEE Conf. Comput. Vis. Pattern Recog., 2022, pp. 9666–9676.
- T. Sosea and C. Caragea, “Marginmatch: Improving semi-supervised learning with pseudo-margins,” in IEEE Conf. Comput. Vis. Pattern Recog. IEEE, 2023, pp. 15 773–15 782.
- H. Chen, R. Tao, Y. Fan, Y. Wang, J. Wang, B. Schiele, X. Xie, B. Raj, and M. Savvides, “Softmatch: Addressing the quantity-quality trade-off in semi-supervised learning,” in Int. Conf. Learn. Represent., 2023.
- D. Berthelot, N. Carlini, I. J. Goodfellow, N. Papernot, A. Oliver, and C. Raffel, “Mixmatch: A holistic approach to semi-supervised learning,” in Adv. Neural Inform. Process. Syst., 2019, pp. 5050–5060.
- E. Malach and S. Shalev-Shwartz, “Decoupling "when to update" from "how to update",” in Adv. Neural Inform. Process. Syst., 2017, pp. 960–970.
- X. Xia, T. Liu, B. Han, C. Gong, N. Wang, Z. Ge, and Y. Chang, “Robust early-learning: Hindering the memorization of noisy labels,” in Int. Conf. Learn. Represent., 2020.
- Z. Sun, H. Liu, Q. Wang, T. Zhou, Q. Wu, and Z. Tang, “Co-ldl: A co-training-based label distribution learning method for tackling label noise,” IEEE Trans. Multimedia, pp. 1093–1104, 2022.
- M. Chen, H. Cheng, Y. Du, M. Xu, W. Jiang, and C. Wang, “Two wrongs don’t make a right: Combating confirmation bias in learning with label noise,” in AAAI, B. Williams, Y. Chen, and J. Neville, Eds., 2023, pp. 14 765–14 773.
- A. Krizhevsky, “Learning multiple layers of features from tiny images,” Technical report, University of Toronto, 2009.
- Z. Sun, X. Hua, Y. Yao, X. Wei, G. Hu, and J. Zhang, “CRSSC: salvage reusable samples from noisy data for robust learning,” in ACM Int. Conf. Multimedia, 2020, pp. 92–101.
- D. Hendrycks, M. Mazeika, S. Kadavath, and D. Song, “Using self-supervised learning can improve model robustness and uncertainty,” in Adv. Neural Inform. Process. Syst., 2019, pp. 15 637–15 648.
- D. Mandal, S. Bharadwaj, and S. Biswas, “A novel self-supervised re-labeling approach for training with noisy labels,” in IEEE Winter Conference on Applications of Computer Vision, 2020, pp. 1370–1379.
- X. Peng, K. Wang, Z. Zeng, Q. Li, J. Yang, and Y. Qiao, “Suppressing mislabeled data via grouping and self-attention,” in Eur. Conf. Comput. Vis., 2020, pp. 786–802.
- L. Huang, C. Zhang, and H. Zhang, “Self-adaptive training: beyond empirical risk minimization,” in Adv. Neural Inform. Process. Syst., 2020.
- Z. Sun, Y. Yao, X. Wei, Y. Zhang, F. Shen, J. Wu, J. Zhang, and H. T. Shen, “Webly supervised fine-grained recognition: Benchmark datasets and an approach,” in Int. Conf. Comput. Vis., 2021, pp. 10 582–10 591.
- S. Liu, Z. Zhu, Q. Qu, and C. You, “Robust training under label noise by over-parameterization,” in Int. Conf. Mach. Learn., 2022, pp. 14 153–14 172.
- X. Shi, Z. Guo, K. Li, Y. Liang, and X. Zhu, “Self-paced resistance learning against overfitting on noisy labels,” Pattern Recognition, vol. 134, p. 109080, 2023.
- E. D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q. V. Le, “Autoaugment: Learning augmentation strategies from data,” in IEEE Conf. Comput. Vis. Pattern Recog., 2019, pp. 113–123.
- Y. Kim, J. Yun, H. Shon, and J. Kim, “Joint negative and positive learning for noisy labels,” in IEEE Conf. Comput. Vis. Pattern Recog., 2021, pp. 9442–9451.
- Huafeng Liu (29 papers)
- Mengmeng Sheng (4 papers)
- Zeren Sun (13 papers)
- Yazhou Yao (52 papers)
- Xian-Sheng Hua (85 papers)
- Heng-Tao Shen (8 papers)