Reducing Bias in Pre-trained Models by Tuning while Penalizing Change (2404.12292v1)
Abstract: Deep models trained on large amounts of data often incorporate implicit biases present during training time. If later such a bias is discovered during inference or deployment, it is often necessary to acquire new data and retrain the model. This behavior is especially problematic in critical areas such as autonomous driving or medical decision-making. In these scenarios, new data is often expensive and hard to come by. In this work, we present a method based on change penalization that takes a pre-trained model and adapts the weights to mitigate a previously detected bias. We achieve this by tuning a zero-initialized copy of a frozen pre-trained network. Our method needs very few, in extreme cases only a single, examples that contradict the bias to increase performance. Additionally, we propose an early stopping criterion to modify baselines and reduce overfitting. We evaluate our approach on a well-known bias in skin lesion classification and three other datasets from the domain shift literature. We find that our approach works especially well with very few images. Simple fine-tuning combined with our early stopping also leads to performance benefits for a larger number of tuning samples.
- International skin imaging collaboration, ISIC Archive. https://www.isic-archive.com/.
- Memory aware synapses: Learning what (not) to forget. In Proceedings of the European conference on computer vision (ECCV), pages 139–154.
- From detection of individual metastases to classification of lymph node status at the patient level: the camelyon17 challenge. IEEE transactions on medical imaging, 38(2):550–560.
- Regularization techniques for fine-tuning in neural machine translation. arXiv preprint arXiv:1707.09920.
- The balanced accuracy and its posterior distribution. In 2010 20th International Conference on Pattern Recognition, pages 3121–3124.
- Adaptation of maximum entropy capitalizer: Little data can help a lot. Computer Speech & Language, 20(4):382–399.
- Proxymix: Proxy-based mixup training with label refinery for source-free domain adaptation. arXiv preprint arXiv:2205.14566.
- Debiasing pre-trained language models via efficient fine-tuning. In Proceedings of the Second Workshop on Language Technology for Equality, Diversity and Inclusion, pages 59–69.
- Deep Learning. MIT Press. http://www.deeplearningbook.org.
- Distance-based regularisation of deep networks for fine-tuning. In ICLR.
- Deep Residual Learning for Image Recognition. arXiv:1512.03385 [cs]. arXiv: 1512.03385.
- Evaluation and mitigation of racial bias in clinical machine learning models: Scoping review. JMIR Med. Inform., 10(5):e36388.
- Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International conference on machine learning, pages 448–456. PMLR.
- Wilds: A benchmark of in-the-wild distribution shifts. In International Conference on Machine Learning, pages 5637–5664. PMLR.
- Feature alignment by uncertainty and self-training for source-free unsupervised domain adaptation. Neural Networks, 161:682–692.
- Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation. In International conference on machine learning, pages 6028–6039. PMLR.
- Deep learning face attributes in the wild. In Proceedings of International Conference on Computer Vision (ICCV).
- An overview of melanoma detection in dermoscopy images using image processing and machine learning. arXiv preprint arXiv:1601.07843.
- Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32.
- Transfer of learning. International encyclopedia of education, 2:6452–6457.
- Bmd: A general class-balanced multicentric dynamic prototype strategy for source-free domain adaptation. In European Conference on Computer Vision, pages 165–182. Springer.
- Conditional adversarial debiasing: Towards learning unbiased classifiers from biased data. In DAGM German Conference on Pattern Recognition (DAGM-GCPR), pages 48–62.
- Interpretations are useful: penalizing explanations to align neural networks with prior knowledge. In International conference on machine learning, pages 8116–8126. PMLR.
- Fairbatch: Batch selection for model fairness. In International Conference on Learning Representations.
- ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 115(3):211–252.
- Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization. In International Conference on Learning Representations.
- Intra-processing methods for debiasing neural networks. Advances in Neural Information Processing Systems, 33:2798–2810.
- The study of nevi in children: Principles learned and implications for melanoma diagnosis. J. Am. Acad. Dermatol., 75(4):813–823.
- End: Entangling and disentangling deep representations for bias correction. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 13508–13517.
- Vdm-da: Virtual domain modeling for source data-free domain adaptation. IEEE Transactions on Circuits and Systems for Video Technology, 32(6):3749–3760.
- The Caltech-UCSD Birds-200-2011 Dataset. Technical Report CNS-TR-2011-001, California Institute of Technology.
- Explicit inductive bias for transfer learning with convolutional networks. In International Conference on Machine Learning, pages 2825–2834. PMLR.
- On early stopping in gradient descent learning. Constructive Approximation, 26:289–315.
- A comprehensive survey on source-free domain adaptation. arXiv preprint arXiv:2302.11803.
- Side-tuning: a baseline for network adaptation via additive side networks. In European Conference on Computer Vision, pages 698–714. Springer.
- Adding conditional control to text-to-image diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3836–3847.
- Places: A 10 million image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence.
- A comprehensive survey on transfer learning. Proceedings of the IEEE, 109(1):43–76.