Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Tackling Noisy Labels with Network Parameter Additive Decomposition (2403.13241v2)

Published 20 Mar 2024 in cs.LG

Abstract: Given data with noisy labels, over-parameterized deep networks suffer overfitting mislabeled data, resulting in poor generalization. The memorization effect of deep networks shows that although the networks have the ability to memorize all noisy data, they would first memorize clean training data, and then gradually memorize mislabeled training data. A simple and effective method that exploits the memorization effect to combat noisy labels is early stopping. However, early stopping cannot distinguish the memorization of clean data and mislabeled data, resulting in the network still inevitably overfitting mislabeled data in the early training stage.In this paper, to decouple the memorization of clean data and mislabeled data, and further reduce the side effect of mislabeled data, we perform additive decomposition on network parameters. Namely, all parameters are additively decomposed into two groups, i.e., parameters $\mathbf{w}$ are decomposed as $\mathbf{w}=\bm{\sigma}+\bm{\gamma}$. Afterward, the parameters $\bm{\sigma}$ are considered to memorize clean data, while the parameters $\bm{\gamma}$ are considered to memorize mislabeled data. Benefiting from the memorization effect, the updates of the parameters $\bm{\sigma}$ are encouraged to fully memorize clean data in early training, and then discouraged with the increase of training epochs to reduce interference of mislabeled data. The updates of the parameters $\bm{\gamma}$ are the opposite. In testing, only the parameters $\bm{\sigma}$ are employed to enhance generalization. Extensive experiments on both simulated and real-world benchmarks confirm the superior performance of our method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (82)
  1. Deep learning. nature, 521(7553):436–444, 2015.
  2. Co-mining: Deep face recognition with noisy labels. In ICCV, pages 9358–9367, 2019.
  3. Combinatorial inference against label noise. In NeurIPS, 2019.
  4. Collaborative refining for person re-identification with label noise. IEEE Transactions on Image Processing, 31:379–391, 2021.
  5. Large-scale pre-training for person re-identification with noisy labels. In CVPR, pages 2476–2486, 2022.
  6. Adaptive hierarchical similarity metric learning with noisy labels. IEEE Transactions on Image Processing, 32:1245–1256, 2023.
  7. Moderate coreset: A universal method of data selection for real-world data-efficient deep learning. In ICLR, 2023.
  8. Ideal: Influence-driven selective annotations empower in-context learners in large language models. arXiv preprint arXiv:2310.10873, 2023.
  9. Error-bounded correction of noisy labels. In ICML, pages 11447–11457, 2020.
  10. Dividemix: Learning with noisy labels as semi-supervised learning. In ICLR, 2020.
  11. Class2simi: A noise reduction perspective on learning with noisy labels. In ICML, 2021.
  12. Searching to exploit memorization effect in learning with noisy labels. In ICML, pages 10789–10798, 2020.
  13. Decode: Deep confidence network for robust image classification. IEEE Transactions on Image Processing, 28(8):3752–3765, 2019.
  14. Unsupervised label noise modeling and loss correction. In ICML, pages 312–321, 2019.
  15. Holistic label correction for noisy multi-label classification. In ICCV, pages 1483–1493, 2023.
  16. Understanding deep learning requires rethinking generalization. In ICLR, 2017.
  17. Can cross entropy loss be robust to label noise? In IJCAI, pages 2206–2212, 2021.
  18. Curriculumnet: Weakly supervised learning from large-scale web images. In ECCV, pages 135–150, 2018.
  19. Selfie: Refurbishing unclean samples for robust deep learning. In ICML, pages 5907–5915, 2019.
  20. Metalabelnet: Learning to generate soft-labels from noisy-labels. IEEE Transactions on Image Processing, 31:4352–4362, 2022.
  21. Lr-svm+: Learning using privileged information with noisy labels. IEEE Transactions on Multimedia, 24:1080–1092, 2021.
  22. Noise against noise: stochastic label noise helps combat inherent label noise. In ICLR, 2021.
  23. Sigua: Forgetting may make learning with noisy labels more robust. In ICML, 2020.
  24. Estimating noise transition matrix with label correlations for noisy multi-label learning. In NeurIPS, pages 24184–24198, 2022.
  25. A closer look at memorization in deep networks. In ICML, pages 233–242, 2017.
  26. Early-learning regularization prevents memorization of noisy labels. In NeurIPS, 2020.
  27. Gradient descent with early stopping is provably robust to label noise for overparameterized neural networks. In AISTATS, 2020.
  28. Simple and effective regularization methods for training on noisily labeled data with generalization guarantee. In ICLR, 2020.
  29. How does early stopping help generalization against label noise? arXiv preprint arXiv:1911.08059, 2019.
  30. Scalable and order-robust continual learning with additive parameter decomposition. arXiv preprint arXiv:1902.09432, 2019.
  31. A survey of label-noise representation learning: Past, present and future. arXiv preprint arXiv:2011.04406, 2020.
  32. Deep learning from noisy image labels with quality embedding. IEEE Transactions on Image Processing, 28(4):1909–1922, 2018.
  33. MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In ICML, pages 2309–2318, 2018.
  34. Learning to reweight examples for robust deep learning. In ICML, pages 4331–4340, 2018.
  35. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In NeurIPS, pages 8527–8537, 2018.
  36. How does disagreement benefit co-teaching? In ICML, 2019.
  37. Combating noisy labels by agreement: A joint training method with co-regularization. In CVPR, pages 13726–13735, 2020.
  38. Combating noisy labels with sample selection by mining high-discrepancy examples. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1833–1843, 2023.
  39. Arash Vahdat. Toward robustness against label noise in training deep discriminative neural networks. In NeurIPS, 2017.
  40. Learning from noisy labels with distillation. In ICCV, pages 1910–1918, 2017.
  41. Joint optimization framework for learning with noisy labels. In CVPR, 2018.
  42. Probabilistic end-to-end noise correction for learning with noisy labels. In CVPR, pages 7017–7025, 2019.
  43. Beyond class-conditional assumption: A primary attempt to combat instance-dependent label noise. arXiv preprint arXiv:2012.05458, 2020.
  44. Robust early-learning: Hindering the memorization of noisy labels. In ICLR, 2021.
  45. A holistic view of label noise transition matrix in deep learning and beyond. In ICLR, 2023.
  46. Classification with noisy labels by importance reweighting. IEEE Transactions on pattern analysis and machine intelligence, 38(3):447–461, 2016.
  47. Training deep neural-networks using a noise adaptation layer. In ICLR, 2017.
  48. Making deep neural networks robust to label noise: A loss correction approach. In CVPR, 2017.
  49. Are anchor points really indispensable in label-noise learning? In NeurIPS, 2019.
  50. Dual t: Reducing estimation error for transition matrix in label-noise learning. In NeurIPS, 2020.
  51. Extended T: Learning with mixed closed-set and open-set noisy labels. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  52. A second-order approach to learning with instance-dependent label noise. In CVPR, 2021.
  53. Meta transition adaptation for robust deep learning with noisy labels. arXiv preprint arXiv:2006.05697, 2020.
  54. Leveraging inter-rater agreement for classification in the presence of noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3439–3448, 2023.
  55. Identifiability of label noise transition matrix. In International Conference on Machine Learning, pages 21475–21496. PMLR, 2023.
  56. Robust loss functions under label noise for deep neural networks. In AAAI, 2017.
  57. Generalized cross entropy loss for training deep neural networks with noisy labels. In NeurIPS, pages 8778–8788, 2018.
  58. Symmetric cross entropy for robust learning with noisy labels. In ICCV, pages 322–330, 2019.
  59. Normalized loss functions for deep learning with noisy labels. In ICML, pages 6543–6553, 2020.
  60. Curriculum loss: Robust learning and generalization against label corruption. In ICLR, 2020.
  61. Mitigating memorization of noisy labels by clipping the model prediction. In International Conference on Machine Learning, pages 36868–36886. PMLR, 2023.
  62. Improve noise tolerance of robust loss via noise-awareness. arXiv preprint arXiv:2301.07306, 2023.
  63. The MNIST database of handwritten digits.
  64. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747, 2017.
  65. Alex Krizhevsky. Learning multiple layers of features from tiny images. Technical report, 2009.
  66. Selective-supervised contrastive learning with noisy labels. In CVPR, pages 316–325, 2022.
  67. Learning with noisy labels via sparse regularization. In ICCV, pages 72–81, 2021.
  68. Nlnl: Negative learning for noisy labels. In ICCV, pages 101–110, 2019.
  69. Coresets for robust training of neural networks against noisy labels. In NeurIPS, 2020.
  70. Scalable penalized regression for noise detection in learning with noisy labels. In CVPR, pages 346–355, 2022.
  71. Learning from noisy labels by regularized estimation of annotator confusion. In CVPR, pages 11244–11253, 2019.
  72. Deep self-learning from noisy labels. In ICCV, pages 5138–5147, 2019.
  73. Part-dependent label noise: Towards instance-dependent label noise. In NeurIPS, 2020.
  74. Food-101–mining discriminative components with random forests. In ECCV, pages 446–461, 2014.
  75. Learning from massive noisy labeled data for image classification. In CVPR, pages 2691–2699, 2015.
  76. Learning with noisy labels revisited: A study using real-world human annotations. In ICLR, 2022.
  77. Sample selection with uncertainty of losses for learning with noisy labels. In ICLR, 2022.
  78. Self: Learning to filter noisy labels with self-ensembling. In ICLR, 2020.
  79. Wide residual networks. arXiv preprint arXiv:1605.07146, 2016.
  80. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
  81. Aggregated residual transformations for deep neural networks. IEEE, 2016.
  82. Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
Citations (5)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets