Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

When Neural Networks Fail to Generalize? A Model Sensitivity Perspective (2212.00850v1)

Published 1 Dec 2022 in cs.CV and cs.AI

Abstract: Domain generalization (DG) aims to train a model to perform well in unseen domains under different distributions. This paper considers a more realistic yet more challenging scenario,namely Single Domain Generalization (Single-DG), where only a single source domain is available for training. To tackle this challenge, we first try to understand when neural networks fail to generalize? We empirically ascertain a property of a model that correlates strongly with its generalization that we coin as "model sensitivity". Based on our analysis, we propose a novel strategy of Spectral Adversarial Data Augmentation (SADA) to generate augmented images targeted at the highly sensitive frequencies. Models trained with these hard-to-learn samples can effectively suppress the sensitivity in the frequency space, which leads to improved generalization performance. Extensive experiments on multiple public datasets demonstrate the superiority of our approach, which surpasses the state-of-the-art single-DG methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Jiajin Zhang (18 papers)
  2. Hanqing Chao (18 papers)
  3. Amit Dhurandhar (62 papers)
  4. Pin-Yu Chen (311 papers)
  5. Ali Tajer (49 papers)
  6. Yangyang Xu (80 papers)
  7. Pingkun Yan (55 papers)
Citations (14)

Summary

We haven't generated a summary for this paper yet.