Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Invariant Information Bottleneck for Domain Generalization (2106.06333v6)

Published 11 Jun 2021 in cs.LG and stat.ML

Abstract: Invariant risk minimization (IRM) has recently emerged as a promising alternative for domain generalization. Nevertheless, the loss function is difficult to optimize for nonlinear classifiers and the original optimization objective could fail when pseudo-invariant features and geometric skews exist. Inspired by IRM, in this paper we propose a novel formulation for domain generalization, dubbed invariant information bottleneck (IIB). IIB aims at minimizing invariant risks for nonlinear classifiers and simultaneously mitigating the impact of pseudo-invariant features and geometric skews. Specifically, we first present a novel formulation for invariant causal prediction via mutual information. Then we adopt the variational formulation of the mutual information to develop a tractable loss function for nonlinear classifiers. To overcome the failure modes of IRM, we propose to minimize the mutual information between the inputs and the corresponding representations. IIB significantly outperforms IRM on synthetic datasets, where the pseudo-invariant features and geometric skews occur, showing the effectiveness of proposed formulation in overcoming failure modes of IRM. Furthermore, experiments on DomainBed show that IIB outperforms $13$ baselines by $0.9\%$ on average across $7$ real datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Bo Li (1107 papers)
  2. Yifei Shen (55 papers)
  3. Yezhen Wang (13 papers)
  4. Wenzhen Zhu (2 papers)
  5. Colorado J. Reed (9 papers)
  6. Jun Zhang (1008 papers)
  7. Dongsheng Li (240 papers)
  8. Kurt Keutzer (200 papers)
  9. Han Zhao (159 papers)
Citations (93)

Summary

We haven't generated a summary for this paper yet.