A Neurosymbolic Framework for Bias Correction in Convolutional Neural Networks (2405.15886v4)
Abstract: Recent efforts in interpreting Convolutional Neural Networks (CNNs) focus on translating the activation of CNN filters into a stratified Answer Set Program (ASP) rule-sets. The CNN filters are known to capture high-level image concepts, thus the predicates in the rule-set are mapped to the concept that their corresponding filter represents. Hence, the rule-set exemplifies the decision-making process of the CNN w.r.t the concepts that it learns for any image classification task. These rule-sets help understand the biases in CNNs, although correcting the biases remains a challenge. We introduce a neurosymbolic framework called NeSyBiCor for bias correction in a trained CNN. Given symbolic concepts, as ASP constraints, that the CNN is biased towards, we convert the concepts to their corresponding vector representations. Then, the CNN is retrained using our novel semantic similarity loss that pushes the filters away from (or towards) learning the desired/undesired concepts. The final ASP rule-set obtained after retraining, satisfies the constraints to a high degree, thus showing the revision in the knowledge of the CNN. We demonstrate that our NeSyBiCor framework successfully corrects the biases of CNNs trained with subsets of classes from the "Places" dataset while sacrificing minimal accuracy and improving interpretability.
- Integrating state-of-the-art cnns for multi-sensor 3d vehicle detection in real autonomous driving environments. In Proc. ITSC 2019, pp. 1425–1431. IEEE.
- A convex framework for fair regression.
- Analysis of explainers of black box deep neural networks for computer vision: A survey.
- Gender shades: Intersectional accuracy disparities in commercial gender classification. In Proc. FAccT 2018, pp. 77–91. PMLR.
- Mitigating reporting bias in semi-supervised temporal commonsense inference with probabilistic soft logic. In Proc. AAAI 2022, pp. 10454–10462.
- Multiwinner voting with fairness constraints. In Proc. IJCAI 2018, pp. 144–151.
- Bias in machine learning software: Why? how? what to do? In Proc. ACM ESEC/FSE 2021, pp. 429–440.
- Imagenet: A large-scale hierarchical image database. In Proc. CVPR 2009, pp. 248–255. Ieee.
- Knowledge representation, reasoning, and the design of intelligent agents: The answer-set programming approach. Cambridge University Press.
- Satisfying real-world goals with dataset constraints. In Proc. NeurIPS 2016, pp. 2415–2423.
- Adam: A method for stochastic optimization. In Proc. ICLR 2015.
- Human evaluation of models built for interpretability. In Proc. HCOMP 2019, volume 7, pp. 59–67.
- Backpropagation applied to handwritten zip code recognition.
- Self-correcting neural networks for safe classification. In Proc. NSV 2022, pp. 96–130. Springer.
- Recognition by association via learning per-exemplar distances. In Proc. CVPR 2008, pp. 1–8.
- A survey on bias and fairness in machine learning.
- Patchresnet: Multiple patch division-based deep feature fusion framework for brain tumor classification using MRI images.
- Nesyfold: A framework for interpretable image classification. In Proc. AAAI 2024a, pp. 4378–4387. AAAI Press.
- Using logic programming and kernel-grouping for improving interpretability of convolutional neural networks. In Proc. PADL 2024b, volume 14512 of LNCS, pp. 134–150. Springer.
- ”why should i trust you?”: Explaining the predictions of any classifier.
- Logic of differentiable logics: Towards a uniform semantics of dl. In 24th International Conference on Logic for Programming, Artificial Intelligence and Reasoning, LPAR 2023 2023, pp. 473–493.
- Unbiased look at dataset bias. In Proc. CVPR 2011, pp. 1521–1528. IEEE.
- Eric: Extracting relations inferred from convolutions. In Computer Vision – ACCV 2021, pp. 206–222, Cham. Springer International.
- FOLD-SE: an efficient rule-based machine learning algorithm with scalable explainability. In Proc. PADL 2024, volume 14512 of LNCS, pp. 37–53. Springer.
- Seggpt: Segmenting everything in context.
- Fairness constraints: Mechanisms for fair classification. In Artificial intelligence and statistics 2017, pp. 962–970.
- Interpreting cnn knowledge via an explanatory graph. In Proc. AAAI 2018, volume 32.
- Growing interpretable part graphs on convnets via multi-shot learning. In Proceedings of the AAAI Conference on Artificial Intelligence 2017, volume 31.
- Men also like shopping: Reducing gender bias amplification using corpus-level constraints. In Proc. EMNLP 2017.
- Places: A 10 million image database for scene recognition.
- Scene parsing through ade20k dataset. In Proc. IEEE CVPR 2017b.