Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Robust Classifiers with Self-Guided Spurious Correlation Mitigation (2405.03649v1)

Published 6 May 2024 in cs.LG and cs.CV

Abstract: Deep neural classifiers tend to rely on spurious correlations between spurious attributes of inputs and targets to make predictions, which could jeopardize their generalization capability. Training classifiers robust to spurious correlations typically relies on annotations of spurious correlations in data, which are often expensive to get. In this paper, we tackle an annotation-free setting and propose a self-guided spurious correlation mitigation framework. Our framework automatically constructs fine-grained training labels tailored for a classifier obtained with empirical risk minimization to improve its robustness against spurious correlations. The fine-grained training labels are formulated with different prediction behaviors of the classifier identified in a novel spuriousness embedding space. We construct the space with automatically detected conceptual attributes and a novel spuriousness metric which measures how likely a class-attribute correlation is exploited for predictions. We demonstrate that training the classifier to distinguish different prediction behaviors reduces its reliance on spurious correlations without knowing them a priori and outperforms prior methods on five real-world datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (48)
  1. Meaningfully debugging model mistakes using conceptual counterfactual explanations. In International Conference on Machine Learning, pages 66–88. PMLR, 2022.
  2. Invariant risk minimization. arXiv preprint arXiv:1907.02893, 2019.
  3. Masktune: Mitigating spurious correlations by forcing to explore. NeurIPS, 35:23284–23296, 2022.
  4. Learning de-biased representations with biased representations. In International Conference on Machine Learning, pages 528–539. PMLR, 2020.
  5. Recognition in terra incognita. In Proceedings of the European conference on computer vision (ECCV), pages 456–473, 2018.
  6. What is the effect of importance weighting in deep learning? In International conference on machine learning, pages 872–881. PMLR, 2019.
  7. Rubi: Reducing unimodal biases for visual question answering. NeurIPS, 32, 2019.
  8. Don’t take the easy way out: Ensemble based methods for avoiding known dataset biases. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4069–4082, 2019.
  9. Environment inference for invariant learning. In International Conference on Machine Learning, pages 2189–2200. PMLR, 2021.
  10. Class-balanced loss based on effective number of samples. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9268–9277, 2019.
  11. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 248–255, 2009.
  12. Robust learning with progressive data expansion against spurious correlation. arXiv preprint arXiv:2306.04949, 2023.
  13. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2020.
  14. Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. arXiv preprint arXiv:1811.12231, 2018.
  15. Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In ICLR, 2019.
  16. Shortcut learning in deep neural networks. Nature Machine Intelligence, 2(11):665–673, 2020.
  17. Umix: Improving importance weighting for subpopulation shift via uncertainty-aware mixup. NeurIPS, 35:37704–37718, 2022.
  18. Learning from imbalanced data. IEEE Transactions on knowledge and data engineering, 21(9):1263–1284, 2009.
  19. Towards non-iid image classification: A dataset and baselines. Pattern Recognition, 110:107383, 2021.
  20. Natural adversarial examples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15262–15271, 2021.
  21. On feature learning in the presence of spurious correlations. NeurIPS, 35:38516–38532, 2022.
  22. Learning debiased classifier with biased committee. NeurIPS, 35:18403–18415, 2022.
  23. Last layer re-training is sufficient for robustness to spurious correlations. In The Eleventh ICLR, 2023.
  24. Diversify and disambiguate: Out-of-distribution robustness via disagreement. In The Eleventh ICLR, 2022.
  25. Large-scale methods for distributionally robust optimization. NeurIPS, 33:8847–8860, 2020.
  26. Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision, pages 3730–3738, 2015.
  27. Just train twice: Improving group robustness without training group information. In International Conference on Machine Learning, pages 6781–6792. PMLR, 2021.
  28. Learning from failure: De-biasing classifier from biased classifier. NeurIPS, 33:20673–20684, 2020.
  29. Spread spurious attribute: Improving worst-group accuracy with spurious attribute estimation. In ICLR, 2022.
  30. Uncovering and correcting shortcut learning in machine learning models for skin cancer diagnosis. Diagnostics, 12(1):40, 2021.
  31. Spurious features everywhere–large-scale detection of harmful spurious features in imagenet. arXiv preprint arXiv:2212.04871, 2022.
  32. NLP Connect. vit-gpt2-image-captioning (revision 0e334c7), 2022.
  33. Towards accountable ai: Hybrid human-machine analyses for characterizing system failure. In Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, volume 6, pages 126–135, 2018.
  34. Finding and fixing spurious patterns with explanations. Transactions on Machine Learning Research, 2022. Expert Certification.
  35. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
  36. Distributionally robust neural networks. In ICLR, 2019.
  37. An investigation of why overparameterization exacerbates spurious correlations. In International Conference on Machine Learning, pages 8346–8356. PMLR, 2020.
  38. Salient imagenet: How to discover spurious features in deep learning? In ICLR, 2021.
  39. Causal attention for unbiased visual recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3091–3100, 2021.
  40. Caltech-UCSD Birds 200. Technical Report CNS-TR-2010-001, California Institute of Technology, 2010.
  41. Cbam: Convolutional block attention module. In Proceedings of the European conference on computer vision (ECCV), pages 3–19, 2018.
  42. Discover and cure: Concept-aware mitigation of spurious correlation. arXiv preprint arXiv:2305.00650, 2023.
  43. Noise or signal: The role of image backgrounds in object recognition. In ICLR, 2021.
  44. Improving out-of-distribution robustness via selective augmentation. In International Conference on Machine Learning, pages 25407–25437. PMLR, 2022.
  45. mixup: Beyond empirical risk minimization. In ICLR, 2018.
  46. Manifold: A model-agnostic framework for interpretation and diagnosis of machine learning models. IEEE transactions on visualization and computer graphics, 25(1):364–373, 2018.
  47. Correct-n-contrast: a contrastive approach for improving robustness to spurious correlations. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato, editors, Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 26484–26516. PMLR, 17–23 Jul 2022.
  48. Places: A 10 million image database for scene recognition. IEEE transactions on pattern analysis and machine intelligence, 40(6):1452–1464, 2017.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Guangtao Zheng (11 papers)
  2. Wenqian Ye (24 papers)
  3. Aidong Zhang (49 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.