Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unsupervised Concept Discovery Mitigates Spurious Correlations (2402.13368v2)

Published 20 Feb 2024 in cs.LG and cs.CV

Abstract: Models prone to spurious correlations in training data often produce brittle predictions and introduce unintended biases. Addressing this challenge typically involves methods relying on prior knowledge and group annotation to remove spurious correlations, which may not be readily available in many applications. In this paper, we establish a novel connection between unsupervised object-centric learning and mitigation of spurious correlations. Instead of directly inferring subgroups with varying correlations with labels, our approach focuses on discovering concepts: discrete ideas that are shared across input samples. Leveraging existing object-centric representation learning, we introduce CoBalT: a concept balancing technique that effectively mitigates spurious correlations without requiring human labeling of subgroups. Evaluation across the benchmark datasets for sub-population shifts demonstrate superior or competitive performance compared state-of-the-art baselines, without the need for group annotation. Code is available at https://github.com/rarefin/CoBalT.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (42)
  1. Invariant risk minimization. arXiv preprint arXiv:1907.02893, 2019.
  2. Masktune: Mitigating spurious correlations by forcing to explore. In Advances in Neural Information Processing Systems, 2022.
  3. Why do deep convolutional networks generalize so poorly to small image transformations? Journal of Machine Learning Research, 20(184):1–25, 2019. URL http://jmlr.org/papers/v20/19-519.html.
  4. Recognition in terra incognita. In Proceedings of the European Conference on Computer Vision (ECCV), September 2018.
  5. Unsupervised learning of visual features by contrasting cluster assignments. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems, volume 33, pp.  9912–9924. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/70feb62b69f16e0238f741fab228fec2-Paper.pdf.
  6. Emerging properties in self-supervised vision transformers. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp.  9630–9640, 2021. URL https://api.semanticscholar.org/CorpusID:233444273.
  7. A simple framework for contrastive learning of visual representations. In III, H. D. and Singh, A. (eds.), Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pp.  1597–1607. PMLR, 13–18 Jul 2020. URL https://proceedings.mlr.press/v119/chen20j.html.
  8. Picie: Unsupervised semantic segmentation using invariance and equivariance in clustering. In CVPR, 2021.
  9. Genesis-v2: Inferring unordered object representations without iterative refinement. In Neural Information Processing Systems, 2021. URL https://api.semanticscholar.org/CorpusID:233307216.
  10. Shortcut learning in deep neural networks. Nature Machine Intelligence, 2:665–673, 2020.
  11. Bootstrap your own latent - a new approach to self-supervised learning. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems, volume 33, pp.  21271–21284. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/f3ada80d5c4ee70142b17b8192b2958e-Paper.pdf.
  12. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  770–778, 2016.
  13. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  9729–9738, 2020.
  14. Simple data balancing achieves competitive worst-group-accuracy. In Schölkopf, B., Uhler, C., and Zhang, K. (eds.), Proceedings of the First Conference on Causal Learning and Reasoning, volume 177 of Proceedings of Machine Learning Research, pp.  336–351. PMLR, 11–13 Apr 2022. URL https://proceedings.mlr.press/v177/idrissi22a.html.
  15. Sgd on neural networks learns functions of increasing complexity. Advances in neural information processing systems, 32, 2019.
  16. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), San Diega, CA, USA, 2015.
  17. Last layer re-training is sufficient for robustness to spurious correlations. 2023. URL https://openreview.net/forum?id=Zb6c8A-Fghk.
  18. Just train twice: Improving group robustness without training group information. In International Conference on Machine Learning, pp. 6781–6792. PMLR, 2021.
  19. Deep learning face attributes in the wild. 2015 IEEE International Conference on Computer Vision (ICCV), pp.  3730–3738, 2014.
  20. Object-centric learning with slot attention. Advances in Neural Information Processing Systems, 33:11525–11538, 2020.
  21. Spread spurious attribute: Improving worst-group accuracy with spurious attribute estimation. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=_F9xpOrqyX9.
  22. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
  23. Discovering environments with xrm. arXiv:2309.16748 [cs.LG], 2023.
  24. Simple and fast group robustness by automatic feature reweighting. In Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., and Scarlett, J. (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp.  28448–28467. PMLR, 23–29 Jul 2023. URL https://proceedings.mlr.press/v202/qiu23c.html.
  25. Rousseeuw, P. J. Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. Journal of computational and applied mathematics, 20:53–65, 1987.
  26. Towards a better understanding of vector quantized autoencoders. 2018.
  27. Distributionally robust neural networks. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum?id=ryxGuJrFvS.
  28. Counterfactual generative networks. ArXiv, abs/2101.06046, 2021. URL https://api.semanticscholar.org/CorpusID:231627872.
  29. Bridging the gap to real-world object-centric learning. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=b9tUk-f_aG.
  30. The pitfalls of simplicity bias in neural networks. Advances in Neural Information Processing Systems, 33:9573–9585, 2020.
  31. No subclass left behind: Fine-grained robustness in coarse-grained classification problems. Advances in Neural Information Processing Systems, 33:19339–19352, 2020.
  32. Robust representation learning via perceptual similarity metrics. In International Conference on Machine Learning, pp. 10043–10053. PMLR, 2021.
  33. Group robust classification without any group information. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=2OcNWFHFpk.
  34. Neural discrete representation learning. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R. (eds.), Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL https://proceedings.neurips.cc/paper_files/paper/2017/file/7a98af17e63a0ac09ce2e96d03992fbc-Paper.pdf.
  35. Self-supervised visual representation learning with semantic grouping. Advances in Neural Information Processing Systems, 35:16423–16438, 2022.
  36. Noise or signal: The role of image backgrounds in object recognition. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=gl3D-xY7wLq.
  37. Identifying spurious biases early in training through the lens of simplicity bias. arXiv preprint arXiv:2305.18761, 2023a.
  38. Change is hard: A closer look at subpopulation shift. In International Conference on Machine Learning, 2023b.
  39. ZIN: When and how to learn invariance without environment partition? In Oh, A. H., Agarwal, A., Belgrave, D., and Cho, K. (eds.), Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=pUPFRSxfACD.
  40. Barlow twins: Self-supervised learning via redundancy reduction. In International Conference on Machine Learning, pp. 12310–12320. PMLR, 2021.
  41. Correct-n-contrast: a contrastive approach for improving robustness to spurious correlations, 2022. URL https://openreview.net/forum?id=cVak2hs06z.
  42. Unlocking slot attention by changing optimal transport costs. In International Conference on machine Learning, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Md Rifat Arefin (11 papers)
  2. Yan Zhang (954 papers)
  3. Aristide Baratin (26 papers)
  4. Francesco Locatello (92 papers)
  5. Irina Rish (85 papers)
  6. Dianbo Liu (59 papers)
  7. Kenji Kawaguchi (147 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com