Out-of-Domain Robustness via Targeted Augmentations (2302.11861v3)
Abstract: Models trained on one set of domains often suffer performance drops on unseen domains, e.g., when wildlife monitoring models are deployed in new camera locations. In this work, we study principles for designing data augmentations for out-of-domain (OOD) generalization. In particular, we focus on real-world scenarios in which some domain-dependent features are robust, i.e., some features that vary across domains are predictive OOD. For example, in the wildlife monitoring application above, image backgrounds vary across camera locations but indicate habitat type, which helps predict the species of photographed animals. Motivated by theoretical analysis on a linear setting, we propose targeted augmentations, which selectively randomize spurious domain-dependent features while preserving robust ones. We prove that targeted augmentations improve OOD performance, allowing models to generalize better with fewer domains. In contrast, existing approaches such as generic augmentations, which fail to randomize domain-dependent features, and domain-invariant augmentations, which randomize all domain-dependent features, both perform poorly OOD. In experiments on three real-world datasets, we show that targeted augmentations set new states-of-the-art for OOD performance by 3.2-15.2 percentage points.
- Adversarial invariant feature learning with accuracy constraint for domain generalization. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 315–331. Springer, 2020.
- Generalizing to unseen domains via distribution matching. arXiv preprint arXiv:1911.00804, 2019.
- Invariant risk minimization. arXiv preprint arXiv:1907.02893, 2019.
- From detection of individual metastases to classification of lymph node status at the patient level: the camelyon17 challenge. IEEE transactions on medical imaging, 38(2):550–560, 2018.
- Recognition in terra incognita. In Proceedings of the European conference on computer vision (ECCV), pp. 456–473, 2018.
- Efficient pipeline for camera trap image review. arXiv preprint arXiv:1907.06772, 2019.
- Synthetic examples improve generalization for rare classes. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 863–873, 2020.
- The iwildcam 2021 competition dataset. arXiv preprint arXiv:2105.03494, 2021.
- The surprising impact of mask-head architecture on novel class segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 7015–7025, 2021.
- Generalization bounds for meta-learning: An information-theoretic analysis. Advances in Neural Information Processing Systems, 34:25878–25890, 2021a.
- A group-theoretic framework for data augmentation. The Journal of Machine Learning Research, 21(1):9885–9955, 2020.
- Iterative feature matching: Toward provable domain generalization with logarithmic environments. arXiv preprint arXiv:2106.09913, 2021b.
- Randaugment: Practical automated data augmentation with a reduced search space. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp. 702–703, 2020.
- Underspecification presents challenges for credibility in modern machine learning. Journal of Machine Learning Research, 2020.
- A kernel theory of modern data augmentation. In International Conference on Machine Learning, pp. 1528–1537. PMLR, 2019.
- Improving bird classification with unsupervised sound separation. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 636–640. IEEE, 2022.
- Improved regularization of convolutional neural networks with cutout. arXiv preprint arXiv:1708.04552, 2017.
- Domain-adversarial training of neural networks. The journal of machine learning research, 17(1):2096–2030, 2016.
- Affinity and diversity: Quantifying mechanisms of data augmentation. arXiv preprint arXiv:2002.08973, 2020.
- In search of lost domain generalization. arXiv preprint arXiv:2007.01434, 2020.
- A distribution-free theory of nonparametric regression, volume 1. Springer, 2002.
- Data augmentation revisited: Rethinking the distribution gap between clean and augmented data. arXiv preprint arXiv:1909.09148, 2019.
- The many faces of robustness: A critical analysis of out-of-distribution generalization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8340–8349, 2021.
- Cycada: Cycle-consistent adversarial domain adaptation. In International conference on machine learning, pp. 1989–1998. Pmlr, 2018.
- A collection of fully-annotated soundscape recordings from the Southwestern Amazon Basin, September 2022. URL https://doi.org/10.5281/zenodo.7079124.
- An analysis of random design linear regression. arXiv preprint arXiv:1106.2363, 2011.
- Selecting data augmentation for simulating interventions. In International Conference on Machine Learning, pp. 4555–4562. PMLR, 2021.
- Support and invertibility in domain-invariant representations. In The 22nd International Conference on Artificial Intelligence and Statistics, pp. 527–536. PMLR, 2019.
- Overview of lifeclef 2021: an evaluation of machine-learning based species identification and species distribution prediction. In Experimental IR Meets Multilinguality, Multimodality, and Interaction: 12th International Conference of the CLEF Association, CLEF 2021, Virtual Event, September 21–24, 2021, Proceedings, pp. 371–393. Springer, 2021.
- Information-theoretic generalization bounds for meta-learning and applications. Entropy, 23(1):126, 2021.
- A collection of fully-annotated soundscape recordings from the Northeastern United States, August 2022. URL https://doi.org/10.5281/zenodo.7079380.
- Wilds: A benchmark of in-the-wild distribution shifts. In International Conference on Machine Learning, pp. 5637–5664. PMLR, 2021.
- How to fine-tune vision models with sgd. arXiv preprint arXiv:2211.09359, 2022.
- Conditional adversarial domain adaptation. Advances in neural information processing systems, 31, 2018.
- On the benefits of invariance in neural networks. arXiv preprint arXiv:2005.00178, 2020.
- Domain generalization using causal matching. In International Conference on Machine Learning, pp. 7313–7324. PMLR, 2021.
- Accuracy on the line: on the strong correlation between out-of-distribution and in-distribution generalization. In International Conference on Machine Learning, pp. 7721–7735. PMLR, 2021.
- A collection of fully-annotated soundscape recordings from the Island of Hawai’i, September 2022. URL https://doi.org/10.5281/zenodo.7078499.
- Specaugment: A simple data augmentation method for automatic speech recognition. arXiv preprint arXiv:1904.08779, 2019.
- Nuisances via negativa: Adjusting for spurious correlations via data augmentation. arXiv preprint arXiv:2210.01302, 2022.
- Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748–8763. PMLR, 2021.
- Model-based domain generalization. Advances in Neural Information Processing Systems, 34:20210–20229, 2021.
- The risks of invariant risk minimization. arXiv preprint arXiv:2010.05761, 2020.
- Quantification of histochemical staining by color deconvolution. Analytical and quantitative cytology and histology, 23(4):291–299, 2001.
- Extending the wilds benchmark for unsupervised adaptation. arXiv preprint arXiv:2112.05090, 2021.
- Sainburg, T. Noise reduction in python using spectral gating, 2022. URL https://github.com/timsainb/noisereduce.
- Deep coral: Correlation alignment for deep domain adaptation. In European conference on computer vision, pp. 443–450. Springer, 2016.
- Correlation alignment for unsupervised domain adaptation. In Domain Adaptation in Computer Vision Applications, pp. 153–171. Springer, 2017.
- Whole-slide mitosis detection in h&e breast histology using phh3 as a reference to train distilled stain-invariant convolutional networks. IEEE transactions on medical imaging, 37(9):2126–2136, 2018.
- Quantifying the effects of data augmentation and stain color normalization in convolutional neural networks for computational pathology. Medical image analysis, 58, 2019.
- Out-of-distribution generalization with causal invariant transformations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 375–385, 2022.
- Heterogeneous domain generalization via domain mixup. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3622–3626. IEEE, 2020.
- A fine-grained analysis on distribution shift. In International Conference on Learning Representations, 2021.
- Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time. In International Conference on Machine Learning, pp. 23965–23998. PMLR, 2022.
- Adversarial domain adaptation with domain mixup. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp. 6502–6509, 2020.
- Improve unsupervised domain adaptation with mixup training. arXiv preprint arXiv:2001.00677, 2020.
- Improving out-of-distribution robustness via selective augmentation. arXiv preprint arXiv:2201.00299, 2022.
- Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 6023–6032, 2019.
- mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412, 2017.
- On learning invariant representations for domain adaptation. In International Conference on Machine Learning, pp. 7523–7532. PMLR, 2019.
- Deep domain-adversarial image generation for domain generalisation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp. 13025–13032, 2020a.
- Learning to generate novel domains for domain generalization. In European conference on computer vision, pp. 561–578. Springer, 2020b.
- Zhu, S. A short note on the tail bound of wishart distribution. arXiv preprint arXiv:1212.5860, 2012.
- Irena Gao (10 papers)
- Shiori Sagawa (12 papers)
- Pang Wei Koh (64 papers)
- Tatsunori Hashimoto (80 papers)
- Percy Liang (239 papers)