Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 172 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 38 tok/s Pro
GPT-5 High 30 tok/s Pro
GPT-4o 73 tok/s Pro
Kimi K2 231 tok/s Pro
GPT OSS 120B 427 tok/s Pro
Claude Sonnet 4.5 38 tok/s Pro
2000 character limit reached

Towards Calibrated Deep Clustering Network (2403.02998v3)

Published 4 Mar 2024 in cs.CV

Abstract: Deep clustering has exhibited remarkable performance; however, the over confidence problem, i.e., the estimated confidence for a sample belonging to a particular cluster greatly exceeds its actual prediction accuracy, has been over looked in prior research. To tackle this critical issue, we pioneer the development of a calibrated deep clustering framework. Specifically, we propose a novel dual head (calibration head and clustering head) deep clustering model that can effectively calibrate the estimated confidence and the actual accuracy. The calibration head adjusts the overconfident predictions of the clustering head, generating prediction confidence that matches the model learning status. Then, the clustering head dynamically selects reliable high-confidence samples estimated by the calibration head for pseudo-label self-training. Additionally, we introduce an effective network initialization strategy that enhances both training speed and network robustness. The effectiveness of the proposed calibration approach and initialization strategy are both endorsed with solid theoretical guarantees. Extensive experiments demonstrate the proposed calibrated deep clustering model not only surpasses the state-of-the-art deep clustering methods by 5x on average in terms of expected calibration error, but also significantly outperforms them in terms of clustering accuracy. The code is available at https://github.com/ChengJianH/CDC.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. Unsupervised neural networks for automatic arabic text summarization using document clustering and topic modeling. Expert Systems with Applications, 172:114652, 2021.
  2. Deep adaptive image clustering. In ICCV, pages 5880–5888, 2017.
  3. A simple framework for contrastive learning of visual representations. In ICML, pages 1597–1607, 2020a.
  4. Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297, 2020b.
  5. An analysis of single-layer networks in unsupervised feature learning. In International Conference on Artificial Intelligence and Statistics, pages 215–223, 2011.
  6. Imagenet: A large-scale hierarchical image database. In CVPR, pages 248–255, 2009.
  7. Local temperature scaling for probability calibration. In ICCV, pages 6889–6899, 2021.
  8. Can we trust you? on calibration of a probabilistic object detector for autonomous driving. arXiv preprint arXiv:1909.12358, 2019.
  9. A survey of uncertainty in deep neural networks. arXiv preprint arXiv:2107.03342, 2021.
  10. Bootstrap your own latent-a new approach to self-supervised learning. NeurIPS, 33:21271–21284, 2020.
  11. On calibration of modern neural networks. In ICML, pages 1321–1330, 2017a.
  12. Improved deep embedded clustering with local structure preservation. In IJCAI, pages 1753–1759, 2017b.
  13. Deep embedding network for clustering. In 2014 International Conference on Pattern Recognition, pages 1532–1537, 2014.
  14. Learning representation for clustering via prototype scattering and positive sampling. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  15. Comparing partitions. Journal of classification, 2:193–218, 1985.
  16. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In ICML, pages 448–456, 2015.
  17. Revisiting explicit regularization in neural networks for well-calibrated predictive uncertainty. arXiv preprint arXiv:2006.06399, 2020.
  18. Learning multiple layers of features from tiny images. 2009.
  19. Verified uncertainty calibration. NeurIPS, 32, 2019.
  20. Tiny imagenet visual recognition challenge. CS 231N, 7(7):3, 2015.
  21. The relationships among various nonnegative matrix factorization methods for clustering. In ICDM, pages 362–371, 2006.
  22. Contrastive clustering. In AAAI, pages 8547–8555, 2021.
  23. Stuart Lloyd. Least squares quantization in pcm. IEEE Transactions on Information Theory, 28(2):129–137, 1982.
  24. Diagnostic uncertainty calibration: Towards reliable machine predictions in medical domain. In International Conference on Artificial Intelligence and Statistics, pages 3664–3672, 2021.
  25. Calibrating deep neural networks using focal loss. NeurIPS, 33:15288–15299, 2020.
  26. When does label smoothing help? NeurIPS, 32, 2019.
  27. Obtaining well calibrated probabilities using bayesian binning. In AAAI, 2015.
  28. Rectified linear units improve restricted boltzmann machines. In ICML, pages 807–814, 2010.
  29. Spice: Semantic pseudo-labeling for image clustering. IEEE Transactions on Image Processing, 31:7264–7278, 2022.
  30. Deep video action clustering via spatio-temporal feature learning. Neurocomputing, 456:519–527, 2021.
  31. Regularizing neural networks by penalizing confident output distributions. arXiv preprint arXiv:1701.06548, 2017.
  32. Qi Qian. Stable cluster discrimination for deep clustering. In CVPR, 2023.
  33. Learning for single-shot confidence calibration in deep neural networks through stochastic inferences. In CVPR, pages 9030–9038, 2019.
  34. You never cluster alone. NeurIPS, 34:27734–27746, 2021.
  35. Cluster ensembles—a knowledge reuse framework for combining multiple partitions. Journal of machine learning research, 3(Dec):583–617, 2002.
  36. Learning deep representations for graph clustering. In AAAI, 2014.
  37. Scan: Learning to classify images without labels. In ECCV, pages 268–285, 2020.
  38. Unsupervised deep embedding for clustering analysis. In ICML, pages 478–487, 2016.
  39. Individual calibration with randomized forecasting. In ICML, pages 11387–11397, 2020.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.