Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

P$^2$OT: Progressive Partial Optimal Transport for Deep Imbalanced Clustering (2401.09266v1)

Published 17 Jan 2024 in cs.CV

Abstract: Deep clustering, which learns representation and semantic clustering without labels information, poses a great challenge for deep learning-based approaches. Despite significant progress in recent years, most existing methods focus on uniformly distributed datasets, significantly limiting the practical applicability of their methods. In this paper, we first introduce a more practical problem setting named deep imbalanced clustering, where the underlying classes exhibit an imbalance distribution. To tackle this problem, we propose a novel pseudo-labeling-based learning framework. Our framework formulates pseudo-label generation as a progressive partial optimal transport problem, which progressively transports each sample to imbalanced clusters under prior distribution constraints, thus generating imbalance-aware pseudo-labels and learning from high-confident samples. In addition, we transform the initial formulation into an unbalanced optimal transport problem with augmented constraints, which can be solved efficiently by a fast matrix scaling algorithm. Experiments on various datasets, including a human-curated long-tailed CIFAR100, challenging ImageNet-R, and large-scale subsets of fine-grained iNaturalist2018 datasets, demonstrate the superiority of our method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (55)
  1. Pseudo-labeling and confirmation bias in deep semi-supervised learning. In 2020 International Joint Conference on Neural Networks (IJCNN), pp.  1–8. IEEE, 2020.
  2. Self-labelling via simultaneous clustering and representation learning. In International Conference on Learning Representations (ICLR), 2020.
  3. Reverse engineering self-supervised learning. arXiv preprint arXiv:2305.15614, 2023.
  4. Free boundaries in optimal transport and monge-ampere obstacle problems. Annals of mathematics, pp.  673–730, 2010.
  5. Learning imbalanced datasets with label-distribution-aware margin loss. Advances in neural information processing systems, 32, 2019.
  6. Deep clustering for unsupervised learning of visual features. In Proceedings of the European conference on computer vision (ECCV), pp.  132–149, 2018.
  7. Unsupervised learning of visual features by contrasting cluster assignments. Advances in Neural Information Processing Systems, 33:9912–9924, 2020.
  8. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pp.  9650–9660, 2021.
  9. Deep adaptive image clustering. In Proceedings of the IEEE international conference on computer vision, pp.  5879–5887, 2017.
  10. Unified optimal transport framework for universal domain adaptation. Advances in Neural Information Processing Systems, 35:29512–29524, 2022.
  11. Csot: Curriculum and structure-aware optimal transport for learning with noisy labels. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  12. Partial optimal tranport with applications on positive-unlabeled learning. Advances in Neural Information Processing Systems, 33:2903–2913, 2020.
  13. Scaling algorithms for unbalanced optimal transport problems. Mathematics of Computation, 87(314):2563–2609, 2018.
  14. Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. Advances in neural information processing systems, 26, 2013.
  15. Nearest neighbor matching for deep clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  13693–13702, June 2021.
  16. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2020.
  17. Optimal transport for domain adaptation. IEEE Trans. Pattern Anal. Mach. Intell, 1:1–40, 2016.
  18. Learning with a wasserstein loss. Advances in neural information processing systems, 28, 2015.
  19. Improved training of wasserstein gans. Advances in neural information processing systems, 30, 2017.
  20. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  16000–16009, 2022.
  21. The many faces of robustness: A critical analysis of out-of-distribution generalization. ICCV, 2021.
  22. Deep semantic clustering by partition confidence maximisation. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
  23. Learning representation for clustering via prototype scattering and positive sampling. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  24. Invariant information clustering for unsupervised image classification and segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  9865–9874, 2019.
  25. Self-damaging contrastive learning. In International Conference on Machine Learning, pp.  4927–4939. PMLR, 2021.
  26. L. Kantorovich. On the transfer of masses (in russian). Doklady Akademii Nauk, 37:227–229, 1942.
  27. Earth movers in the big data era: A review of optimal transport in machine learning. arXiv preprint arXiv:2305.05080, 2023.
  28. Philip A Knight. The sinkhorn–knopp algorithm: convergence and applications. SIAM Journal on Matrix Analysis and Applications, 30(1):261–275, 2008.
  29. Learning multiple layers of features from tiny images. 2009.
  30. Dong-Hyun Lee et al. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on challenges in representation learning, ICML, volume 3, pp.  896. Atlanta, 2013.
  31. Contrastive clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pp.  8547–8555, 2021.
  32. Optimal entropy-transport problems and a new hellinger–kantorovich distance between positive measures. Inventiones mathematicae, 211(3):969–1117, 2018.
  33. Self-supervised learning is more robust to dataset imbalance. 2022.
  34. Cot: Unsupervised domain adaptation with clustering and optimal transport. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  19998–20007, June 2023.
  35. Divclust: Controlling diversity in deep clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  3418–3428, 2023.
  36. Spice: Semantic pseudo-labeling for image clustering. IEEE Transactions on Image Processing, 31:7264–7278, 2022.
  37. Computational optimal transport. Center for Research in Economics and Statistics Working Papers, (2017-86), 2017.
  38. Deepdpm: Deep clustering with an unknown number of clusters. In Conference on Computer Vision and Pattern Recognition, 2022.
  39. Timo Aila Samuli Laine. Temporal ensembling for semi-supervised learning. International Conference on Learning Representations (ICLR), 30, 2017.
  40. You never cluster alone. Advances in Neural Information Processing Systems, 34:27734–27746, 2021.
  41. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. Advances in Neural Information Processing Systems, 33:596–608, 2020.
  42. Transporting labels via hierarchical optimal transport for semi-supervised learning. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IV 16, pp.  509–526. Springer, 2020.
  43. Sinkhorn label allocation: Semi-supervised classification via annealed self-training. In International Conference on Machine Learning, pp.  10065–10075. PMLR, 2021.
  44. Clustering-friendly representation learning via instance discrimination and feature decorrelation. In International Conference on Learning Representations, 2020.
  45. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. Advances in neural information processing systems, 30, 2017.
  46. Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
  47. Scan: Learning to classify images without labels. In Proceedings of the European Conference on Computer Vision, 2020.
  48. The inaturalist species classification and detection dataset. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  8769–8778, 2018.
  49. Cédric Villani et al. Optimal transport: old and new, volume 338. Springer, 2009.
  50. Crest: A class-rebalancing self-training framework for imbalanced semi-supervised learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  10857–10866, June 2021.
  51. Flexmatch: Boosting semi-supervised learning with curriculum pseudo labeling. Advances in Neural Information Processing Systems, 34:18408–18419, 2021a.
  52. Novel class discovery for long-tailed recognition. Transactions on Machine Learning Research, 2023.
  53. Distribution alignment: A unified framework for long-tail visual recognition. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  2361–2370, 2021b.
  54. A comprehensive survey on deep clustering: Taxonomy, challenges, and future directions. arXiv preprint arXiv:2206.07579, 2022a.
  55. Contrastive learning with boosted memorization. In International Conference on Machine Learning, pp.  27367–27377. PMLR, 2022b.
Citations (4)

Summary

We haven't generated a summary for this paper yet.