Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Contrasting the landscape of contrastive and non-contrastive learning (2203.15702v1)

Published 29 Mar 2022 in cs.LG, cs.CV, and stat.ML

Abstract: A lot of recent advances in unsupervised feature learning are based on designing features which are invariant under semantic data augmentations. A common way to do this is contrastive learning, which uses positive and negative samples. Some recent works however have shown promising results for non-contrastive learning, which does not require negative samples. However, the non-contrastive losses have obvious "collapsed" minima, in which the encoders output a constant feature embedding, independent of the input. A folk conjecture is that so long as these collapsed solutions are avoided, the produced feature representations should be good. In our paper, we cast doubt on this story: we show through theoretical results and controlled experiments that even on simple data models, non-contrastive losses have a preponderance of non-collapsed bad minima. Moreover, we show that the training process does not avoid these minima.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Ashwini Pokle (9 papers)
  2. Jinjin Tian (7 papers)
  3. Yuchen Li (85 papers)
  4. Andrej Risteski (58 papers)
Citations (27)

Summary

We haven't generated a summary for this paper yet.