Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multilevel Knowledge Transfer for Cross-Domain Object Detection (2108.00977v2)

Published 2 Aug 2021 in cs.CV and cs.LG

Abstract: Domain shift is a well known problem where a model trained on a particular domain (source) does not perform well when exposed to samples from a different domain (target). Unsupervised methods that can adapt to domain shift are highly desirable as they allow effective utilization of the source data without requiring additional annotated training data from the target. Practically, obtaining sufficient amount of annotated data from the target domain can be both infeasible and extremely expensive. In this work, we address the domain shift problem for the object detection task. Our approach relies on gradually removing the domain shift between the source and the target domains. The key ingredients to our approach are -- (a) mapping the source to the target domain on pixel-level; (b) training a teacher network on the mapped source and the unannotated target domain using adversarial feature alignment; and (c) finally training a student network using the pseudo-labels obtained from the teacher. Experimentally, when tested on challenging scenarios involving domain shift, we consistently obtain significantly large performance gains over various recent state of the art approaches.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Botos Csaba (5 papers)
  2. Xiaojuan Qi (133 papers)
  3. Arslan Chaudhry (15 papers)
  4. Puneet Dokania (4 papers)
  5. Philip Torr (172 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.