Mitigating Shortcut Learning with Diffusion Counterfactuals and Diverse Ensembles (2311.16176v5)
Abstract: Spurious correlations in the data, where multiple cues are predictive of the target labels, often lead to a phenomenon known as shortcut learning, where a model relies on erroneous, easy-to-learn cues while ignoring reliable ones. In this work, we propose DiffDiv an ensemble diversification framework exploiting Diffusion Probabilistic Models (DPMs) to mitigate this form of bias. We show that at particular training intervals, DPMs can generate images with novel feature combinations, even when trained on samples displaying correlated input features. We leverage this crucial property to generate synthetic counterfactuals to increase model diversity via ensemble disagreement. We show that DPM-guided diversification is sufficient to remove dependence on shortcut cues, without a need for additional supervised signals. We further empirically quantify its efficacy on several diversification objectives, and finally show improved generalization and diversification on par with prior work that relies on auxiliary data collection.
- Don’t just assume; look and answer: Overcoming priors for visual question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
- Synthetic data from diffusion models improves imagenet classification. arXiv preprint arXiv:2304.08466, 2023.
- Recognition in terra incognita. In Proceedings of the European Conference on Computer Vision (ECCV), 2018.
- Beyond surface statistics: Scene representations in a latent diffusion model. arXiv preprint arXiv:2306.05720, 2023.
- Shortcut learning in deep neural networks. Nature Machine Intelligence, 2(11):665–673, 2020.
- Learning not to learn: Training deep neural networks with biased data. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9004–9012, Los Alamitos, CA, USA, 2019. IEEE Computer Society.
- Last layer re-training is sufficient for robustness to spurious correlations. In The Eleventh International Conference on Learning Representations, 2023.
- Diffusion models already have a semantic latent space. arXiv preprint arXiv:2210.10960, 2022.
- Diversify and disambiguate: Learning from underspecified data. arXiv preprint arXiv:2202.03418, 2022.
- Repair: Removing representation bias by dataset resampling. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9564–9573, Los Alamitos, CA, USA, 2019. IEEE Computer Society.
- Dpm-solver: A fast ode solver for diffusion probabilistic model sampling in around 10 steps. Advances in Neural Information Processing Systems, 35:5775–5787, 2022.
- dsprites: Disentanglement testing sprites dataset. https://github.com/deepmind/dsprites-dataset/, 2017.
- Automatic shortcut removal for self-supervised representation learning. In Proceedings of the 37th International Conference on Machine Learning. JMLR.org, 2020.
- Learning diverse features in vision transformers for improved generalization. arXiv preprint arXiv:2308.16274, 2023.
- Compositional abilities emerge multiplicatively: Exploring diffusion models on a synthetic task. arXiv preprint arXiv:2310.09336, 2023.
- Agree to disagree: Diversity through disagreement for better transferability. arXiv preprint arXiv:2202.04414, 2022.
- Ensembles of locally independent prediction models. In AAAI, 2020.
- Distributionally robust neural networks. In International Conference on Learning Representations, 2020.
- Progressive distillation for fast sampling of diffusion models. arXiv preprint arXiv:2202.00512, 2022.
- Fake it till you make it: Learning transferable representations from synthetic imagenet clones. In CVPR 2023–IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023.
- Which shortcut cues will dnns choose? a study from the parameter-space perspective. In International Conference on Learning Representations, 2022.
- Leveraging diffusion disentangled representations to mitigate shortcuts in underspecified visual tasks. In NeurIPS 2023 Workshop on Diffusion Models, 2023.
- The pitfalls of simplicity bias in neural networks. In Proceedings of the 34th International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 2020. Curran Associates Inc.
- Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502, 2020.
- Evading the simplicity bias: Training a diverse set of models discovers solutions with superior OOD generalization. In CVPR, 2022a.
- Predicting is not understanding: Recognizing and addressing underspecification in machine learning. arXiv preprint arXiv:2207.02598, 2022b.
- Unbiased look at dataset bias. In CVPR 2011, pages 1521–1528, 2011.
- Balanced datasets are not enough: Estimating and mitigating gender bias in deep image representations. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pages 5309–5318, Los Alamitos, CA, USA, 2019. IEEE Computer Society.
- Concept algebra for (score-based) text-controlled generative models. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
- Uncovering the disentanglement capability in text-to-image diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1900–1910, 2023.
- Noise or signal: The role of image backgrounds in object recognition. In International Conference on Learning Representations, 2021.
- Investigating bias and fairness in facial expression recognition. In Computer Vision – ECCV 2020 Workshops: Glasgow, UK, August 23–28, 2020, Proceedings, Part VI, page 506–523, Berlin, Heidelberg, 2020. Springer-Verlag.
- Not just pretty pictures: Text-to-image generators enable interpretable interventions for robust representations. arXiv preprint arXiv:2212.11237, 2022.
- Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLoS Med., 15(11):e1002683, 2018.
- Age progression/regression by conditional adversarial autoencoder. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5810–5818, 2017.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.