Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Vision Transformer-based Adversarial Domain Adaptation (2404.15817v1)

Published 24 Apr 2024 in cs.CV and cs.LG

Abstract: Unsupervised domain adaptation (UDA) aims to transfer knowledge from a labeled source domain to an unlabeled target domain. The most recent UDA methods always resort to adversarial training to yield state-of-the-art results and a dominant number of existing UDA methods employ convolutional neural networks (CNNs) as feature extractors to learn domain invariant features. Vision transformer (ViT) has attracted tremendous attention since its emergence and has been widely used in various computer vision tasks, such as image classification, object detection, and semantic segmentation, yet its potential in adversarial domain adaptation has never been investigated. In this paper, we fill this gap by employing the ViT as the feature extractor in adversarial domain adaptation. Moreover, we empirically demonstrate that ViT can be a plug-and-play component in adversarial domain adaptation, which means directly replacing the CNN-based feature extractor in existing UDA methods with the ViT-based feature extractor can easily obtain performance improvement. The code is available at https://github.com/LluckyYH/VT-ADA.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Deep learning, MIT press, 2016.
  2. “A survey on vision transformer,” IEEE transactions on pattern analysis and machine intelligence, vol. 45, no. 1, pp. 87–110, 2022.
  3. “A survey on transfer learning,” IEEE Transactions on knowledge and data engineering, vol. 22, no. 10, pp. 1345–1359, 2009.
  4. “Deep visual domain adaptation: A survey,” Neurocomputing, vol. 312, pp. 135–153, 2018.
  5. “Connecting the dots with landmarks: Discriminatively learning domain-invariant features for unsupervised domain adaptation,” in International conference on machine learning. PMLR, 2013, pp. 222–230.
  6. “Mind the class weight bias: Weighted maximum mean discrepancy for unsupervised domain adaptation,” in Proceedings of the CVPR, 2017, pp. 2272–2281.
  7. “Deep coral: Correlation alignment for deep domain adaptation,” in Computer Vision–ECCV 2016 Workshops: Amsterdam, The Netherlands, October 8-10 and 15-16, 2016, Proceedings, Part III 14. Springer, 2016, pp. 443–450.
  8. “Generative adversarial nets,” Advances in neural information processing systems, vol. 27, 2014.
  9. “Domain-adversarial training of neural networks,” The journal of machine learning research, vol. 17, no. 1, pp. 2096–2030, 2016.
  10. “Conditional adversarial domain adaptation,” Advances in neural information processing systems, vol. 31, 2018.
  11. “A survey of convolutional neural networks: analysis, applications, and prospects,” IEEE transactions on neural networks and learning systems, 2021.
  12. “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
  13. “Swin transformer v2: Scaling up capacity and resolution,” in Proceedings of the CVPR, 2022, pp. 12009–12019.
  14. “Exploring plain vision transformer backbones for object detection,” in European Conference on Computer Vision. Springer, 2022, pp. 280–296.
  15. “Segmenter: Transformer for semantic segmentation,” in Proceedings of the ICCV, 2021, pp. 7262–7272.
  16. “A survey on data augmentation in large model era,” arXiv preprint arXiv:2401.15422, 2024.
  17. “Dual mixup regularized learning for adversarial domain adaptation,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXIX 16. Springer, 2020, pp. 540–555.
  18. “Adversarial discriminative domain adaptation,” in Proceedings of the CVPR, 2017, pp. 7167–7176.
  19. “Cdtrans: Cross-domain transformer for unsupervised domain adaptation,” in International Conference on Learning Representations, 2021.
  20. “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
  21. “A survey on evaluation of large language models,” ACM Transactions on Intelligent Systems and Technology, 2023.
  22. “Enhanced transport distance for unsupervised domain adaptation,” in Proceedings of the CVPR, 2020, pp. 13936–13944.
  23. “Unsupervised domain adaptation by backpropagation,” in International conference on machine learning. PMLR, 2015, pp. 1180–1189.
  24. Laurens Van der Maaten and Geoffrey Hinton, “Visualizing data using t-sne.,” Journal of machine learning research, vol. 9, no. 11, 2008.

Summary

We haven't generated a summary for this paper yet.