Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
166 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Preventing Collapse in Contrastive Learning with Orthonormal Prototypes (CLOP) (2403.18699v2)

Published 27 Mar 2024 in cs.LG and cs.AI

Abstract: Contrastive learning has emerged as a powerful method in deep learning, excelling at learning effective representations through contrasting samples from different distributions. However, neural collapse, where embeddings converge into a lower-dimensional space, poses a significant challenge, especially in semi-supervised and self-supervised setups. In this paper, we first theoretically analyze the effect of large learning rates on contrastive losses that solely rely on the cosine similarity metric, and derive a theoretical bound to mitigate this collapse. {Building on these insights, we propose CLOP, a novel semi-supervised loss function designed to prevent neural collapse by promoting the formation of orthogonal linear subspaces among class embeddings.} Unlike prior approaches that enforce a simplex ETF structure, CLOP focuses on subspace separation, leading to more distinguishable embeddings. Through extensive experiments on real and synthetic datasets, we demonstrate that CLOP enhances performance, providing greater stability across different learning rates and batch sizes.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. Vicreg: Variance-invariance-covariance regularization for self-supervised learning. arXiv preprint arXiv:2105.04906, 2021.
  2. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597–1607. PMLR, 2020a.
  3. Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297, 2020b.
  4. Debiased contrastive learning. Advances in neural information processing systems, 33:8765–8775, 2020.
  5. Parametric contrastive learning. In Proceedings of the IEEE/CVF international conference on computer vision, pages 715–724, 2021.
  6. Hyperbolic contrastive learning for visual representations beyond objects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6840–6849, 2023.
  7. Olivier Henaff. Data-efficient image recognition with contrastive predictive coding. In International conference on machine learning, pages 4182–4192. PMLR, 2020.
  8. A survey on contrastive self-supervised learning. Technologies, 9(1):2, 2020.
  9. Geometric contrastive learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 206–215, 2023.
  10. Prototypical contrastive learning of unsupervised representations. arXiv preprint arXiv:2005.04966, 2020.
  11. Contrastive multiview coding. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XI 16, pages 776–794. Springer, 2020a.
  12. What makes for good views for contrastive learning? Advances in neural information processing systems, 33:6827–6839, 2020b.
  13. Rethinking minimal sufficient representation in contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16041–16050, 2022.
  14. Unsupervised feature learning by cross-level instance-group discrimination. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 12586–12595, 2021.
  15. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3733–3742, 2018.
  16. What should not be contrastive in contrastive learning. arXiv preprint arXiv:2008.05659, 2020.
  17. Stable contrastive learning for self-supervised sentence embeddings with pseudo-siamese mutual learning. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 30:3046–3059, 2022.
  18. Decoupled contrastive learning. In European Conference on Computer Vision, pages 668–684. Springer, 2022.
  19. Barlow twins: Self-supervised learning via redundancy reduction. In International Conference on Machine Learning, pages 12310–12320. PMLR, 2021.
  20. Deep robust clustering by contrastive learning. arXiv preprint arXiv:2008.03030, 2020.

Summary

We haven't generated a summary for this paper yet.