Towards Calibrated Deep Clustering Network (2403.02998v3)
Abstract: Deep clustering has exhibited remarkable performance; however, the over confidence problem, i.e., the estimated confidence for a sample belonging to a particular cluster greatly exceeds its actual prediction accuracy, has been over looked in prior research. To tackle this critical issue, we pioneer the development of a calibrated deep clustering framework. Specifically, we propose a novel dual head (calibration head and clustering head) deep clustering model that can effectively calibrate the estimated confidence and the actual accuracy. The calibration head adjusts the overconfident predictions of the clustering head, generating prediction confidence that matches the model learning status. Then, the clustering head dynamically selects reliable high-confidence samples estimated by the calibration head for pseudo-label self-training. Additionally, we introduce an effective network initialization strategy that enhances both training speed and network robustness. The effectiveness of the proposed calibration approach and initialization strategy are both endorsed with solid theoretical guarantees. Extensive experiments demonstrate the proposed calibrated deep clustering model not only surpasses the state-of-the-art deep clustering methods by 5x on average in terms of expected calibration error, but also significantly outperforms them in terms of clustering accuracy. The code is available at https://github.com/ChengJianH/CDC.
- Unsupervised neural networks for automatic arabic text summarization using document clustering and topic modeling. Expert Systems with Applications, 172:114652, 2021.
- Deep adaptive image clustering. In ICCV, pages 5880–5888, 2017.
- A simple framework for contrastive learning of visual representations. In ICML, pages 1597–1607, 2020a.
- Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297, 2020b.
- An analysis of single-layer networks in unsupervised feature learning. In International Conference on Artificial Intelligence and Statistics, pages 215–223, 2011.
- Imagenet: A large-scale hierarchical image database. In CVPR, pages 248–255, 2009.
- Local temperature scaling for probability calibration. In ICCV, pages 6889–6899, 2021.
- Can we trust you? on calibration of a probabilistic object detector for autonomous driving. arXiv preprint arXiv:1909.12358, 2019.
- A survey of uncertainty in deep neural networks. arXiv preprint arXiv:2107.03342, 2021.
- Bootstrap your own latent-a new approach to self-supervised learning. NeurIPS, 33:21271–21284, 2020.
- On calibration of modern neural networks. In ICML, pages 1321–1330, 2017a.
- Improved deep embedded clustering with local structure preservation. In IJCAI, pages 1753–1759, 2017b.
- Deep embedding network for clustering. In 2014 International Conference on Pattern Recognition, pages 1532–1537, 2014.
- Learning representation for clustering via prototype scattering and positive sampling. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
- Comparing partitions. Journal of classification, 2:193–218, 1985.
- Batch normalization: Accelerating deep network training by reducing internal covariate shift. In ICML, pages 448–456, 2015.
- Revisiting explicit regularization in neural networks for well-calibrated predictive uncertainty. arXiv preprint arXiv:2006.06399, 2020.
- Learning multiple layers of features from tiny images. 2009.
- Verified uncertainty calibration. NeurIPS, 32, 2019.
- Tiny imagenet visual recognition challenge. CS 231N, 7(7):3, 2015.
- The relationships among various nonnegative matrix factorization methods for clustering. In ICDM, pages 362–371, 2006.
- Contrastive clustering. In AAAI, pages 8547–8555, 2021.
- Stuart Lloyd. Least squares quantization in pcm. IEEE Transactions on Information Theory, 28(2):129–137, 1982.
- Diagnostic uncertainty calibration: Towards reliable machine predictions in medical domain. In International Conference on Artificial Intelligence and Statistics, pages 3664–3672, 2021.
- Calibrating deep neural networks using focal loss. NeurIPS, 33:15288–15299, 2020.
- When does label smoothing help? NeurIPS, 32, 2019.
- Obtaining well calibrated probabilities using bayesian binning. In AAAI, 2015.
- Rectified linear units improve restricted boltzmann machines. In ICML, pages 807–814, 2010.
- Spice: Semantic pseudo-labeling for image clustering. IEEE Transactions on Image Processing, 31:7264–7278, 2022.
- Deep video action clustering via spatio-temporal feature learning. Neurocomputing, 456:519–527, 2021.
- Regularizing neural networks by penalizing confident output distributions. arXiv preprint arXiv:1701.06548, 2017.
- Qi Qian. Stable cluster discrimination for deep clustering. In CVPR, 2023.
- Learning for single-shot confidence calibration in deep neural networks through stochastic inferences. In CVPR, pages 9030–9038, 2019.
- You never cluster alone. NeurIPS, 34:27734–27746, 2021.
- Cluster ensembles—a knowledge reuse framework for combining multiple partitions. Journal of machine learning research, 3(Dec):583–617, 2002.
- Learning deep representations for graph clustering. In AAAI, 2014.
- Scan: Learning to classify images without labels. In ECCV, pages 268–285, 2020.
- Unsupervised deep embedding for clustering analysis. In ICML, pages 478–487, 2016.
- Individual calibration with randomized forecasting. In ICML, pages 11387–11397, 2020.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.