Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning to Learn Domain-invariant Parameters for Domain Generalization (2211.04582v1)

Published 4 Nov 2022 in cs.LG

Abstract: Due to domain shift, deep neural networks (DNNs) usually fail to generalize well on unknown test data in practice. Domain generalization (DG) aims to overcome this issue by capturing domain-invariant representations from source domains. Motivated by the insight that only partial parameters of DNNs are optimized to extract domain-invariant representations, we expect a general model that is capable of well perceiving and emphatically updating such domain-invariant parameters. In this paper, we propose two modules of Domain Decoupling and Combination (DDC) and Domain-invariance-guided Backpropagation (DIGB), which can encourage such general model to focus on the parameters that have a unified optimization direction between pairs of contrastive samples. Our extensive experiments on two benchmarks have demonstrated that our proposed method has achieved state-of-the-art performance with strong generalization capability.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Feng Hou (14 papers)
  2. Yao Zhang (537 papers)
  3. Yang Liu (2253 papers)
  4. Jin Yuan (22 papers)
  5. Cheng Zhong (30 papers)
  6. Yang Zhang (1129 papers)
  7. Zhongchao Shi (25 papers)
  8. Jianping Fan (51 papers)
  9. Zhiqiang He (37 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.