Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

FroSSL: Frobenius Norm Minimization for Efficient Multiview Self-Supervised Learning (2310.02903v4)

Published 4 Oct 2023 in cs.LG

Abstract: Self-supervised learning (SSL) is a popular paradigm for representation learning. Recent multiview methods can be classified as sample-contrastive, dimension-contrastive, or asymmetric network-based, with each family having its own approach to avoiding informational collapse. While these families converge to solutions of similar quality, it can be empirically shown that some methods are epoch-inefficient and require longer training to reach a target performance. Two main approaches to improving efficiency are covariance eigenvalue regularization and using more views. However, these two approaches are difficult to combine due to the computational complexity of computing eigenvalues. We present the objective function FroSSL which reconciles both approaches while avoiding eigendecomposition entirely. FroSSL works by minimizing covariance Frobenius norms to avoid collapse and minimizing mean-squared error to achieve augmentation invariance. We show that FroSSL reaches competitive accuracies more quickly than any other SSL method and provide theoretical and empirical support that this faster convergence is due to how FroSSL affects the eigenvalues of the embedding covariance matrices. We also show that FroSSL learns competitive representations on linear probe evaluation when used to train a ResNet-18 on several datasets, including STL-10, Tiny ImageNet, and ImageNet-100.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Contrastive and non-contrastive self-supervised learning recover global and local spectral embedding methods. Advances in Neural Information Processing Systems, 35:26671–26685, 2022.
  2. VICReg: Variance-invariance-covariance regularization for self-supervised learning. In International Conference on Learning Representations, 2022.
  3. Unsupervised learning of visual features by contrasting cluster assignments. Advances in Neural Information Processing Systems, 33:9912–9924, 2020.
  4. Emerging properties in self-supervised vision transformers. In IEEE/CVF International Conference on Computer Vision, pp. 9650–9660, 2021.
  5. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp. 1597–1607. PMLR, 2020a.
  6. Exploring simple Siamese representation learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  15750–15758, 2021.
  7. Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297, 2020b.
  8. solo-learn: A library of self-supervised methods for visual representation learning. Journal of Machine Learning Research, 23(56):1–6, 2022. URL http://jmlr.org/papers/v23/21-1155.html.
  9. Whitening for self-supervised representation learning. In International Conference on Machine Learning, pp. 3015–3024. PMLR, 2021.
  10. Rankme: Assessing the downstream performance of pretrained self-supervised representations by their rank. In International Conference on Machine Learning, pp. 10929–10974. PMLR, 2023a.
  11. On the duality between contrastive and non-contrastive self-supervised learning. In International Conference on Learning Representations, 2023b. URL https://openreview.net/forum?id=kDEL91Dufpa.
  12. A kernel statistical test of independence. Advances in Neural Information Processing Systems, 20, 2007.
  13. Bootstrap your own latent-a new approach to self-supervised learning. Advances in Neural Information Processing Systems, 33:21271–21284, 2020.
  14. Provable guarantees for self-supervised deep learning with spectral contrastive loss. Advances in Neural Information Processing Systems, 34:5000–5011, 2021.
  15. Momentum contrast for unsupervised visual representation learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  9729–9738, 2020.
  16. Vne: An effective method for improving deep representation by manipulating eigenvalue distribution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  3799–3810, 2023.
  17. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  18. Sample estimate of the entropy of a random vector. Problemy Peredachi Informatsii, 23(2):9–16, 1987.
  19. Self-supervised learning with kernel dependence maximization. Advances in Neural Information Processing Systems, 34:15543–15556, 2021.
  20. Self-supervised learning with an information maximization criterion. Advances in Neural Information Processing Systems, 35:35240–35253, 2022.
  21. The matrix cookbook. Technical University of Denmark, 7(15):510, 2008.
  22. Measures of entropy from data using infinitely divisible kernels. IEEE Transactions on Information Theory, 61(1):535–548, 2015.
  23. On the stepwise nature of self-supervised learning. In International Conference on Machine Learning, 2023.
  24. Exploring the equivalence of siamese self-supervised learning via a unified gradient framework. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  14431–14440, 2022.
  25. A note on connecting barlow twins with negative-sample-free contrastive learning. arXiv preprint arXiv:2104.13712, 2021a.
  26. Self-supervised learning from a multi-view perspective. In International Conference on Learning Representations, 2021b. URL https://openreview.net/forum?id=-bdp_8Itjwp.
  27. SimMER: Simple maximization of entropy and rank for self-supervised representation learning, 2022. URL https://openreview.net/forum?id=77_zstKV8HQ.
  28. Large batch training of convolutional networks. arXiv preprint arXiv:1708.03888, 2017.
  29. Barlow twins: Self-supervised learning via redundancy reduction. In International Conference on Machine Learning, pp. 12310–12320. PMLR, 2021.
  30. Tico: Transformation invariance and covariance contrast for self-supervised visual representation learning. arXiv preprint arXiv:2206.10698, 2022.

Summary

We haven't generated a summary for this paper yet.