Towards Real World Debiasing: A Fine-grained Analysis On Spurious Correlation (2405.15240v2)
Abstract: Spurious correlations in training data significantly hinder the generalization capability of machine learning models when faced with distribution shifts in real-world scenarios. To tackle the problem, numerous debias approaches have been proposed and benchmarked on datasets intentionally designed with severe biases. However, it remains to be asked: \textit{1. Do existing benchmarks really capture biases in the real world? 2. Can existing debias methods handle biases in the real world?} To answer the questions, we revisit biased distributions in existing benchmarks and real-world datasets, and propose a fine-grained framework for analyzing dataset bias by disentangling it into the magnitude and prevalence of bias. We observe and theoretically demonstrate that existing benchmarks poorly represent real-world biases. We further introduce two novel biased distributions to bridge this gap, forming a nuanced evaluation framework for real-world debiasing. Building upon these results, we evaluate existing debias methods with our evaluation framework. Results show that existing methods are incapable of handling real-world biases. Through in-depth analysis, we propose a simple yet effective approach that can be easily applied to existing debias methods, named Debias in Destruction (DiD). Empirical results demonstrate the superiority of DiD, improving the performance of existing methods on all types of biases within the proposed evaluation framework.
- Computer vision for autonomous vehicles: Problems, datasets and state of the art, 2021.
- The Role of Machine Learning Algorithms for Diagnosing Diseases. Journal of Applied Science and Technology Trends, 2(01):10–19, March 2021. doi: 10.38094/jastt20179.
- WILDS: A benchmark of in-the-Wild distribution shifts. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 5637–5664. PMLR, 2021-07-18/2021-07-24.
- A fine-grained analysis on distribution shift. In International Conference on Learning Representations, 2022.
- Learning debiased representation via disentangled feature augmentation. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 25123–25133. Curran Associates, Inc., 2021.
- Learning from failure: De-biasing classifier from biased classifier. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 20673–20684. Curran Associates, Inc., 2020.
- Learning de-biased representations with biased representations. In Proceedings of the 37th International Conference on Machine Learning, ICML’20. JMLR.org, 2020.
- BiaSwap: Removing dataset bias with bias-tailored swapping augmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 14992–15001, October 2021.
- Learning debiased classifier with biased committee. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors, Advances in Neural Information Processing Systems, volume 35, pages 18403–18415. Curran Associates, Inc., 2022.
- Just train twice: Improving group robustness without training group information. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 6781–6792. PMLR, 2021-07-18/2021-07-24.
- BiasAdv: Bias-Adversarial Augmentation for Model Debiasing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 3832–3841, June 2023.
- Combating Unknown Bias with Effective Bias-Conflicting Scoring and Gradient Alignment. Proceedings of the AAAI Conference on Artificial Intelligence, 37(3):3561–3569, June 2023. doi: 10.1609/aaai.v37i3.25466.
- Benchmarking Bias Mitigation Algorithms in Representation Learning through Fairness Metrics. In J. Vanschoren and S. Yeung, editors, Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks, volume 1, 2021.
- Revisiting the Importance of Amplifying Bias for Debiasing. Proceedings of the AAAI Conference on Artificial Intelligence, 37(12):14974–14981, June 2023. ISSN 2374-3468, 2159-5399. doi: 10.1609/aaai.v37i12.26748.
- Microsoft COCO: Common objects in context, 2015.
- Mitigating gender bias in captioning systems. In Proceedings of the Web Conference 2021, WWW ’21, pages 633–645, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 978-1-4503-8312-7. doi: 10.1145/3442381.3449950.
- Machine Bias. ProPublica, 2016.
- Non-Discriminatory Machine Learning Through Convex Fairness Criteria. Proceedings of the AAAI Conference on Artificial Intelligence, 32(1), April 2018. doi: 10.1609/aaai.v32i1.11662.
- Fairea: A Model Behaviour Mutation Approach to Benchmarking Bias Mitigation Methods. In Proceedings of the 29th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, ESEC/FSE 2021, pages 994–1006, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 978-1-4503-8562-6. doi: 10.1145/3468264.3468565.
- Achieving Fairness at No Utility Cost via Data Reweighing with Influence. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato, editors, Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 12917–12930. PMLR, July 2022.
- Overwriting pretrained bias with finetuning data. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 3957–3968, October 2023.
- FACTS: First amplify correlations and then slice to discover bias. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 4794–4804, October 2023.
- On the foundations of shortcut learning. In The Twelfth International Conference on Learning Representations, 2024.
- Distribution alignment: A unified framework for long-tail visual recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2361–2370, 2021.
- Deep long-tailed learning: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
- Generalized cross entropy loss for training deep neural networks with noisy labels. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS’18, pages 8792–8802, Red Hook, NY, USA, 2018. Curran Associates Inc.
- ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In International Conference on Learning Representations, 2019.
- Peng Kuang (1 paper)
- Zhixuan Chu (43 papers)
- Jingyi Wang (105 papers)
- Kui Ren (169 papers)
- ZhiBo Wang (48 papers)