Improving Label Error Detection and Elimination with Uncertainty Quantification (2405.09602v1)
Abstract: Identifying and handling label errors can significantly enhance the accuracy of supervised machine learning models. Recent approaches for identifying label errors demonstrate that a low self-confidence of models with respect to a certain label represents a good indicator of an erroneous label. However, latest work has built on softmax probabilities to measure self-confidence. In this paper, we argue that -- as softmax probabilities do not reflect a model's predictive uncertainty accurately -- label error detection requires more sophisticated measures of model uncertainty. Therefore, we develop a range of novel, model-agnostic algorithms for Uncertainty Quantification-Based Label Error Detection (UQ-LED), which combine the techniques of confident learning (CL), Monte Carlo Dropout (MCD), model uncertainty measures (e.g., entropy), and ensemble learning to enhance label error detection. We comprehensively evaluate our algorithms on four image classification benchmark datasets in two stages. In the first stage, we demonstrate that our UQ-LED algorithms outperform state-of-the-art confident learning in identifying label errors. In the second stage, we show that removing all identified errors from the training data based on our approach results in higher accuracies than training on all available labeled data. Importantly, besides our contributions to the detection of label errors, we particularly propose a novel approach to generate realistic, class-dependent label errors synthetically. Overall, our study demonstrates that selectively cleaning datasets with UQ-LED algorithms leads to more accurate classifications than using larger, noisier datasets.
- Illustrative Discussion of MC-Dropout in General Dataset: Uncertainty Estimation in Bitcoin. Neural Processing Letters, 53(2), 1001–1011.
- Label Noise Types and Their Effects on Deep Learning. arXiv preprint arXiv:2003.10471, 0(0).
- Learning From Noisy Examples. Machine Learning, 2(4), 343–370.
- Data Excellence for AI: Why Should You Care?. Interactions, 29(2), 66–69.
- Bishop, C. M. (2016). Pattern Recognition and Machine Learning (Softcover reprint of the original 1st edition 2006 (corrected at 8th printing 2009) edition). Information Science and Statistics. Springer New York, New York, NY.
- Learning with Instance-Dependent Label Noise: A Sample Sieve Approach. arXiv preprint arXiv:2010.02347, 0(0).
- Learning with Bounded Instance- and Label-dependent Label Noise. In International Conference on Machine Learning, Vol. 1, pp. 1789–1799. PMLR.
- Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed edition). L. Erlbaum Associates, Hillsdale, N.J.
- A Survey on Deep Learning with Noisy Labels: How to train your model when you cannot trust on the annotations?. In 2020 33rd SIBGRAPI conference on graphics, patterns and images (SIBGRAPI), Vol. 1, pp. 9–16. IEEE.
- Quality Control in Crowdsourcing: A Survey of Quality Attributes, Assessment Techniques and Assurance Actions. ACM Computing Surveys (CSUR), 51(1), 1–40.
- Falcon, W. (2019). PyTorch Lightning..
- Classification in the Presence of Label Noise: A Survey. IEEE Transactions on Neural Networks and Learning Systems, 25(5), 845–869.
- Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In International Conference on Machine Learning, pp. 1050–1059.
- Exploiting Class Similarity for Machine Learning with Confidence Labels and Projective Loss Functions. arXiv preprint arXiv:2103.13607, 0(0).
- Robust Loss Functions under Label Noise for Deep Neural Networks. In Proceedings of the AAAI conference on artificial intelligence, Vol. 31.
- Training deep neural-networks using a noise adaptation layer. In International Conference on Learning Representations.
- AI lifecycle models need to be revised: An exploratory study in Fintech. Empirical Software Engineering, 26(5), 95.
- Hamid, O. H. (2022). From Model-Centric to Data-Centric AI: A Paradigm Shift or Rather a Complementary Approach?. In 2022 8th International Conference on Information Technology Trends (ITT), pp. 196–199, Dubai, United Arab Emirates. IEEE.
- A Survey of Label-noise Representation Learning: Past, Present and Future. arXiv preprint arXiv:2011.04406, 0(0).
- Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels. Advances in neural information processing systems, 31(1).
- Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778.
- A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks. In Proceedings of the International Conference on Learning Representations.
- Using Pre-Training Can Improve Model Robustness and Uncertainty. In International conference on machine learning, pp. 2712–2721. PMLR.
- Hands-on Bayesian Neural Networks – a Tutorial for Deep Learning Users. IEEE Computational Intelligence Magazine, 17(2), 29–48. Publisher: arXiv.
- Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding. arXiv preprint arXiv:1511.02680, 0(0).
- Krizhevsky, A. (2009). Learning Multiple Layers of Features From Tiny Images. Tech. rep., University of Toronto, Toronto.
- Tiny Imagenet Visual Recognition Challenge. Stanford University CS 231N, 7(7), 3.
- MNIST handwritten digit database. ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, 2.
- Does Label Smoothing Mitigate Label Noise?. In International Conference on Machine Learning, pp. 6448–6458. PMLR.
- DataPerf: Benchmarks for Data-Centric AI Development.. Publisher: arXiv.
- Learning with Noisy Labels. In Burges, C. J., Bottou, L., Welling, M., Ghahramani, Z., & Weinberger, K. Q. (Eds.), Advances in Neural Information Processing Systems, Vol. 26. Curran Associates, Inc.
- Neal, R. M. (1996). Bayesian Learning for Neural Networks, Vol. 118 of Lecture Notes in Statistics. Springer New York, New York, NY.
- Pervasive Label Errors in Test Sets Destabilize Machine Learning Benchmarks. In Proceedings of the 35th Conference on Neural Information Processing Systems (NeurIPS 2021) Track on Datasets and Benchmarks.
- Confident Learning: Estimating Uncertainty in Dataset Labels. Journal of Artificial Intelligence Research, 70, 1373–1411.
- Pytorch: An Imperative Style, High-Performance Deep Learning Library. Advances in Neural Information Processing Systems, 32, 8024–8035.
- Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
- “Everyone Wants to Do the Model Work, Not the Data Work”: Data Cascades in High-Stakes AI. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, pp. 1–15.
- Correlation Coefficients: Appropriate Use and Interpretation. Anesthesia & Analgesia, 126(5), 1763–1768.
- Meta-Weight-Net: Learning an Explicit Mapping For Sample Weighting. Advances in neural information processing systems, 32.
- Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations.
- Learning from Noisy Labels with Deep Neural Networks: A Survey.. Publisher: arXiv.
- Dropout: A Simple Way to Prevent Neural Networks from Overfitting. Journal of Machine Learning Research, 15(56), 1929–1958.
- An Uncertainty-Based Human-in-the-Loop System for Industrial Tool Wear Analysis. In Machine Learning and Knowledge Discovery in Databases. Applied Data Science and Demo Track, Vol. 12461, pp. 85–100. Springer International Publishing, Cham. Series Title: Lecture Notes in Computer Science.
- Turney, S. (2022). Pearson Correlation Coefficient (r)..
- Whose AI Dream? In search of the aspiration in data annotation.. In CHI Conference on Human Factors in Computing Systems, pp. 1–16. ACM.
- Crowdsourcing label quality: a theoretical analysis. Science China Information Sciences, 58(11), 1–12.
- Are Anchor Points Really Indispensable in Label-Noise Learning?. Advances in Neural Information Processing Systems, 32.
- Learning from massive noisy labeled data for image classification. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2691–2699, Boston, MA, USA. IEEE.
- Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels. Advances in Neural Information Processing Systems, 31.