Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CLImage: Human-Annotated Datasets for Complementary-Label Learning (2305.08295v3)

Published 15 May 2023 in cs.LG and cs.CV

Abstract: Complementary-label learning (CLL) is a weakly-supervised learning paradigm that aims to train a multi-class classifier using only complementary labels, which indicate classes to which an instance does not belong. Despite numerous algorithmic proposals for CLL, their practical applicability remains unverified for two reasons. Firstly, these algorithms often rely on assumptions about the generation of complementary labels, and it is not clear how far the assumptions are from reality. Secondly, their evaluation has been limited to synthetic datasets. To gain insights into the real-world performance of CLL algorithms, we developed a protocol to collect complementary labels from human annotators. Our efforts resulted in the creation of four datasets: CLCIFAR10, CLCIFAR20, CLMicroImageNet10, and CLMicroImageNet20, derived from well-known classification datasets CIFAR10, CIFAR100, and TinyImageNet200. These datasets represent the very first real-world CLL datasets. Through extensive benchmark experiments, we discovered a notable decrease in performance when transitioning from synthetic datasets to real-world datasets. We investigated the key factors contributing to the decrease with a thorough dataset-level ablation study. Our analyses highlight annotation noise as the most influential factor in the real-world datasets. In addition, we discover that the biased-nature of human-annotated complementary labels and the difficulty to validate with only complementary labels are two outstanding barriers to practical CLL. These findings suggest that the community focus more research efforts on developing CLL algorithms and validation schemes that are robust to noisy and biased complementary-label distributions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (16)
  1. Unbiased risk estimators can mislead: A case study of learning with complementary labels, 2020.
  2. Learning from partial labels. The Journal of Machine Learning Research, 12:1501–1536, 2011.
  3. F. Denis. Pac learning from positive statistical queries. In Algorithmic Learning Theory: 9th International Conference, ALT’98 Otzenhausen, Germany, October 8–10, 1998 Proceedings 9, pages 112–126. Springer, 1998.
  4. Learning with multiple complementary labels. In International Conference on Machine Learning, pages 3072–3081. PMLR, 2020.
  5. B. Frénay and M. Verleysen. Classification in the presence of label noise: a survey. IEEE transactions on neural networks and learning systems, 25(5):845–869, 2013.
  6. Y. Gao and M.-L. Zhang. Discriminative complementary-label learning with weighted loss. In International Conference on Machine Learning, pages 3587–3597. PMLR, 2021.
  7. Learning from complementary labels. Advances in neural information processing systems, 30, 2017.
  8. Complementary-label learning for arbitrary losses and models, 2019.
  9. Learning from noisy complementary labels with robust loss functions. IEICE TRANSACTIONS on Information and Systems, 105(2):364–376, 2022.
  10. Reduction from complementary-label learning to probability estimates. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD), May 2023.
  11. Consistent complementary-label learning via order-preserving losses. In International Conference on Artificial Intelligence and Statistics, pages 8734–8748. PMLR, 2023.
  12. Machine learning from weak supervision: An empirical risk minimization approach. MIT Press, 2022.
  13. Learning from complementary labels via partial-output consistency regularization. In IJCAI, pages 3075–3081, 2021.
  14. Learning with noisy labels revisited: A study using real-world human annotations, 2022.
  15. Learning with biased complementary labels, 2018.
  16. Z.-H. Zhou. A brief introduction to weakly supervised learning. National science review, 5(1):44–53, 2018.
Citations (1)

Summary

We haven't generated a summary for this paper yet.