Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Connect, Not Collapse: Explaining Contrastive Learning for Unsupervised Domain Adaptation (2204.00570v4)

Published 1 Apr 2022 in cs.LG and cs.CV

Abstract: We consider unsupervised domain adaptation (UDA), where labeled data from a source domain (e.g., photographs) and unlabeled data from a target domain (e.g., sketches) are used to learn a classifier for the target domain. Conventional UDA methods (e.g., domain adversarial training) learn domain-invariant features to improve generalization to the target domain. In this paper, we show that contrastive pre-training, which learns features on unlabeled source and target data and then fine-tunes on labeled source data, is competitive with strong UDA methods. However, we find that contrastive pre-training does not learn domain-invariant features, diverging from conventional UDA intuitions. We show theoretically that contrastive pre-training can learn features that vary subtantially across domains but still generalize to the target domain, by disentangling domain and class information. Our results suggest that domain invariance is not necessary for UDA. We empirically validate our theory on benchmark vision datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Kendrick Shen (3 papers)
  2. Robbie Jones (3 papers)
  3. Ananya Kumar (27 papers)
  4. Sang Michael Xie (21 papers)
  5. Jeff Z. HaoChen (12 papers)
  6. Tengyu Ma (117 papers)
  7. Percy Liang (239 papers)
Citations (74)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com