Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

FIXED: Frustratingly Easy Domain Generalization with Mixup (2211.05228v2)

Published 7 Nov 2022 in cs.CV, cs.AI, and cs.LG

Abstract: Domain generalization (DG) aims to learn a generalizable model from multiple training domains such that it can perform well on unseen target domains. A popular strategy is to augment training data to benefit generalization through methods such as Mixup~\cite{zhang2018mixup}. While the vanilla Mixup can be directly applied, theoretical and empirical investigations uncover several shortcomings that limit its performance. Firstly, Mixup cannot effectively identify the domain and class information that can be used for learning invariant representations. Secondly, Mixup may introduce synthetic noisy data points via random interpolation, which lowers its discrimination capability. Based on the analysis, we propose a simple yet effective enhancement for Mixup-based DG, namely domain-invariant Feature mIXup (FIX). It learns domain-invariant representations for Mixup. To further enhance discrimination, we leverage existing techniques to enlarge margins among classes to further propose the domain-invariant Feature MIXup with Enhanced Discrimination (FIXED) approach. We present theoretical insights about guarantees on its effectiveness. Extensive experiments on seven public datasets across two modalities including image classification (Digits-DG, PACS, Office-Home) and time series (DSADS, PAMAP2, UCI-HAR, and USC-HAD) demonstrate that our approach significantly outperforms nine state-of-the-art related methods, beating the best performing baseline by 6.5\% on average in terms of test accuracy. Code is available at: https://github.com/jindongwang/transferlearning/tree/master/code/deep/fixed.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (76)
  1. mixup: Beyond empirical risk minimization. In International Conference on Learning Representations, 2018.
  2. Auto-gait: Automatic ataxia risk assessment with computer vision from gait task videos. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 7(1):1–19, 2023.
  3. Ada Wan. Fairness in representation for multilingual nlp: Insights from controlled experiments on conditional language modeling. In International Conference on Learning Representations, 2021.
  4. Dataset shift in machine learning. Mit Press, 2009.
  5. Analysis of representations for domain adaptation. In Advances in neural information processing systems, volume 19, page 137, 2007.
  6. Domain-adversarial training of neural networks. The journal of machine learning research, 17(1):2096–2030, 2016.
  7. Domain adaptation for statistical classifiers. Journal of artificial Intelligence research, 26:101–126, 2006.
  8. Domain generalization via invariant feature representation. In International Conference on Machine Learning, pages 10–18. PMLR, 2013.
  9. Learning to generalize: Meta-learning for domain generalization. In Thirty-Second AAAI Conference on Artificial Intelligence, 2018a.
  10. Generalizing to unseen domains: A survey on domain generalization. IEEE, 2022.
  11. Domain generalization with adversarial feature learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5400–5409, 2018b.
  12. Metareg: Towards domain generalization using meta-regularization. In Advances in Neural Information Processing Systems, volume 31, pages 998–1008, 2018.
  13. Exploiting domain-specific features to enhance domain generalization. Advances in Neural Information Processing Systems, 34, 2021.
  14. Domain generalization with mixstyle. In International Conference on Learning Representations (ICLR), 2021.
  15. Fsdr: Frequency space domain randomization for domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6891–6902, 2021.
  16. How does mixup help with robustness and generalization? In ICLR, 2021a.
  17. Heterogeneous domain generalization via domain mixup. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 3622–3626. IEEE, 2020a.
  18. A fourier-based framework for domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14383–14392, 2021.
  19. Manifold mixup: Better representations by interpolating hidden states. In International Conference on Machine Learning, pages 6438–6447. PMLR, 2019.
  20. Semantic-discriminative mixup for generalizable sensor-based cross-domain activity recognition. Proceedings of the ACM on Interactive, Mobile, Wearable, and Ubiquitous Technologies, 2022.
  21. Large margin deep networks for classification. In Advances in Neural Information Processing Systems, volume 31, pages 842–852, 2018.
  22. Deep coral: Correlation alignment for deep domain adaptation. In European conference on computer vision, pages 443–450. Springer, 2016.
  23. On mixup regularization. The Journal of Machine Learning Research, 23(1):14632–14662, 2022.
  24. Unsupervised domain adaptation via structurally regularized deep clustering. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 8725–8735, 2020.
  25. Latent dirichlet allocation. the Journal of machine Learning research, 3:993–1022, 2003.
  26. Deep transfer metric learning. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 325–333, 2015.
  27. Selfreg: Self-supervised contrastive regularization for domain generalization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9619–9628, 2021.
  28. Deep domain-adversarial image generation for domain generalisation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 13025–13032, 2020a.
  29. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
  30. Unsupervised domain adaptation by backpropagation. In International conference on machine learning, pages 1180–1189. PMLR, 2015.
  31. Reading digits in natural images with unsupervised feature learning. 2011.
  32. Deeper, broader and artier domain generalization. In Proceedings of the IEEE international conference on computer vision, pages 5542–5550, 2017.
  33. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5018–5027, 2017.
  34. Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation. In International Conference on Machine Learning, pages 6028–6039. PMLR, 2020.
  35. In search of lost domain generalization. In International Conference on Learning Representations, 2021.
  36. Domain generalization by solving jigsaw puzzles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2229–2238, 2019.
  37. Episodic training for domain generalization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1446–1455, 2019.
  38. Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization. In International Conference on Learning Representations (ICLR), 2020.
  39. Self-challenging improves cross-domain generalization. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part II 16, pages 124–140. Springer, 2020.
  40. Learning explanations that are hard to vary. In ICLR, 2021.
  41. Learning to generate novel domains for domain generalization. In European Conference on Computer Vision, pages 561–578. Springer, 2020b.
  42. Reducing domain gap by reducing style bias. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8690–8699, 2021.
  43. Domain generalization using a mixture of multiple latent domains. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 11749–11756, 2020.
  44. Style normalization and restitution for generalizable person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3143–3152, 2020.
  45. Learning from extrinsic and intrinsic supervisions for domain generalization. In European Conference on Computer Vision, pages 159–176. Springer, 2020b.
  46. Efficient domain generalization via common-specific low-rank decomposition. In International Conference on Machine Learning, pages 7728–7738. PMLR, 2020.
  47. Informative dropout for robust representation learning: A shape-bias perspective. In International Conference on Machine Learning, pages 8828–8839. PMLR, 2020.
  48. Towards recognizing unseen categories in unseen domains. In European Conference on Computer Vision, pages 466–483. Springer, 2020.
  49. Deep stable learning for out-of-distribution generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5372–5382, 2021b.
  50. Domain generalization using causal matching. In International Conference on Machine Learning, pages 7313–7324. PMLR, 2021.
  51. Learning to diversify for single domain generalization. pages 834–843, 2021.
  52. A simple feature augmentation for domain generalization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 8886–8895, 2021.
  53. Recognizing daily and sports activities in two open source machine learning environments using body-worn sensor units. The Computer Journal, 57(11):1649–1667, 2014.
  54. Mi Zhang and Alexander A Sawchuk. Usc-had: a daily activity dataset for ubiquitous activity recognition using wearable sensors. In Proceedings of the 2012 ACM conference on ubiquitous computing, pages 1036–1043, 2012.
  55. Human activity recognition on smartphones using a multiclass hardware-friendly support vector machine. In International workshop on ambient assisted living, pages 216–223. Springer, 2012.
  56. Introducing a new benchmarked dataset for activity monitoring. In 2012 16th international symposium on wearable computers, pages 108–109. IEEE, 2012.
  57. Latent independent excitation for generalizable sensor-based cross-person activity recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 11921–11929, 2021.
  58. Deep domain generalization via conditional invariant adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV), pages 624–639, 2018c.
  59. Domain agnostic learning with disentangled representations. In International Conference on Machine Learning, pages 5102–5112. PMLR, 2019.
  60. Domain generalization via model-agnostic learning of semantic features. volume 32, pages 6450–6461, 2019.
  61. Gradient matching for domain generalization. In International Conference on Learning Representations, 2022.
  62. Domain generalization using shape representation. In European Conference on Computer Vision, pages 666–670. Springer, 2020.
  63. Learning to learn single domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12556–12565, 2020.
  64. Improving out-of-distribution robustness via selective augmentation. In International Conference on Machine Learning, pages 25407–25437. PMLR, 2022.
  65. Invariant risk minimization. arXiv preprint arXiv:1907.02893, 2019.
  66. On learning invariant representations for domain adaptation. In International Conference on Machine Learning, pages 7523–7532. PMLR, 2019.
  67. Domain generalization via gradient surgery. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6630–6638, 2021.
  68. Domain randomization and pyramid consistency: Simulation-to-real generalization without accessing target domain data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2100–2110, 2019.
  69. Generalizing across domains via cross-gradient training. In International Conference on Learning Representations, 2018.
  70. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6023–6032, 2019.
  71. Puzzle mix: Exploiting saliency and local statistics for optimal mixup. In International Conference on Machine Learning, pages 5275–5285. PMLR, 2020.
  72. Adversarial domain adaptation with domain mixup. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 6502–6509, 2020.
  73. Dual mixup regularized learning for adversarial domain adaptation. In European Conference on Computer Vision, pages 540–555. Springer, 2020.
  74. A theory of learning from different domains. Machine learning, 79(1):151–175, 2010.
  75. Domain adversarial neural networks for domain generalization: When it works and how to improve. Machine Learning, pages 1–37, 2023.
  76. Adversarial target-invariant representation learning for domain generalization. arXiv preprint arxiv:1911.00804, 2020.
Citations (4)

Summary

  • The paper’s main contribution is FIXED, which applies Mixup on domain-invariant features to enhance generalization across unseen domains.
  • It integrates a large margin loss to clearly separate classes and mitigate noisy feature interpolation, yielding an average 6.5% accuracy improvement.
  • The work provides theoretical insights and robust empirical validation on seven benchmarks, outperforming nine state-of-the-art domain generalization methods.

An Analysis of "FIXED: Frustratingly Easy Domain Generalization with Mixup"

The paper "FIXED: Frustratingly Easy Domain Generalization with Mixup" addresses a critical challenge in machine learning known as Domain Generalization (DG). The core objective of DG is to develop models that perform robustly on unseen domains by learning from multiple training domains. This research specifically enhances the Mixup technique to improve the generalization capabilities of models across different domains.

Key Contributions and Methodology

The primary contribution of the paper is the introduction of the domain-invariant Feature MIXup with Enhanced Discrimination (FIXED). This approach builds on the existing Mixup method to overcome identified limitations:

  1. Domain-Invariant Feature Mixup (FIX): Traditional Mixup methods struggle to separate domain-specific information from class information, leading to the potential entanglement of features that may harm performance. FIX addresses this by performing Mixup on domain-invariant features instead of raw inputs, allowing the model to focus on relevant classification information without domain interference.
  2. Enhanced Discrimination with Large Margin Loss: To mitigate the issue of generating noisy synthetic data points, FIXED incorporates a large margin loss that increases the separation between classes. This addition serves to enhance the discrimination power of the classifiers, ensuring that interpolated features maintain their class-specific characteristics.
  3. Theoretical Insights: The paper provides theoretical justification for the superiority of FIXED over traditional Mixup by analyzing distribution coverage and inter-class distances. It demonstrates that FIXED effectively reduces the risk of generating unrecognizable synthetic data, a major concern with the vanilla Mixup method.

Experimental Results

The experimental evaluation conducted on seven benchmark datasets across two modalities—image classification and time series—demonstrates that FIXED outperforms nine state-of-the-art DG methods, achieving an average improvement of 6.5% on test accuracy. This extensive evaluation underlines the practical efficacy of FIXED in handling diverse and challenging domain shifts.

Implications and Future Directions

The introduction of FIXED has both theoretical and practical implications. Theoretically, it emphasizes the importance of disentangling domain and class information when using data augmentation techniques like Mixup. Practically, FIXED offers a straightforward yet effective enhancement to the standard data augmentation pipeline, making it applicable to a broad range of classification tasks.

Future developments could explore the integration of FIXED with other domain-invariant learning frameworks to further boost generalization performance. Additionally, adapting FIXED for application in regression or non-classification tasks might expand its applicability.

Conclusion

The research presented in "FIXED: Frustratingly Easy Domain Generalization with Mixup" contributes significantly to advancements in domain generalization by refining Mixup for better generalization across unseen domains. The proposed enhancements are backed by theoretical insights and robust empirical results, making FIXED a compelling approach for enhancing model generalization in diverse domains.