Not All Neuro-Symbolic Concepts Are Created Equal: Analysis and Mitigation of Reasoning Shortcuts (2305.19951v2)
Abstract: Neuro-Symbolic (NeSy) predictive models hold the promise of improved compliance with given constraints, systematic generalization, and interpretability, as they allow to infer labels that are consistent with some prior knowledge by reasoning over high-level concepts extracted from sub-symbolic inputs. It was recently shown that NeSy predictors are affected by reasoning shortcuts: they can attain high accuracy but by leveraging concepts with unintended semantics, thus coming short of their promised advantages. Yet, a systematic characterization of reasoning shortcuts and of potential mitigation strategies is missing. This work fills this gap by characterizing them as unintended optima of the learning objective and identifying four key conditions behind their occurrence. Based on this, we derive several natural mitigation strategies, and analyze their efficacy both theoretically and empirically. Our analysis shows reasoning shortcuts are difficult to deal with, casting doubts on the trustworthiness and interpretability of existing NeSy solutions.
- From statistical relational to neural-symbolic artificial intelligence. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, pages 4943–4950, 2021.
- Neural-symbolic learning and reasoning: A survey and interpretation. Neuro-Symbolic Artificial Intelligence: The State of the Art, 342:1, 2022.
- Deep learning with logical constraints. arXiv preprint arXiv:2205.00523, 2022.
- A review of some techniques for inclusion of domain-knowledge into deep neural networks. Scientific Reports, 12(1):1–15, 2022.
- Semantic-based regularization for learning and inference. Artificial Intelligence, 2017.
- Logic tensor networks for semantic image interpretation. In IJCAI, 2017.
- DeepProbLog: Neural Probabilistic Logic Programming. NeurIPS, 2018.
- A semantic loss function for deep learning with symbolic knowledge. In ICML, 2018.
- Coherent hierarchical multi-label classification networks. NeurIPS, 2020.
- Semantic Probabilistic Layers for Neuro-Symbolic Learning. In NeurIPS, 2022a.
- Learning with logical constraints but without shortcut satisfaction. In The Eleventh International Conference on Learning Representations, 2023.
- Neuro-symbolic verification of deep neural networks. 2022.
- Cynthia Rudin. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5):206–215, 2019.
- This looks like that: Deep learning for interpretable image recognition. NeurIPS, 2019.
- Concept whitening for interpretable image recognition. Nature Machine Intelligence, 2020.
- Ai for radiographic covid-19 detection selects shortcuts over signal. Nature Machine Intelligence, pages 1–10, 2021.
- A weakly supervised strategy for learning object detection on a humanoid robot. In 2019 IEEE-RAS 19th International Conference on Humanoid Robots (Humanoids), pages 194–201. IEEE, 2019.
- Self-driving cars: A survey. Expert Systems with Applications, 165:113816, 2021.
- Learning modulo theories. arXiv preprint arXiv:2301.11435, 2023.
- Neural probabilistic logic programming in deepproblog. Artificial Intelligence, 298:103504, 2021a.
- Neuro symbolic continual learning: Knowledge, reasoning shortcuts and concept rehearsal. In ICML, 2023.
- Yann LeCun. The mnist database of handwritten digits. http://yann. lecun. com/exdb/mnist/, 1998.
- Dl2: Training and querying neural networks with logic. In International Conference on Machine Learning, pages 1931–1941. PMLR, 2019.
- Neuro-symbolic entropy regularization. In UAI, 2022b.
- Multiplexnet: Towards fully satisfied logical constraints in neural networks. In AAAI, 2022.
- Bridging logic and kernel machines. Machine learning, 86(1):57–88, 2012.
- Neupsl: Neural probabilistic soft logic. arXiv preprint arXiv:2205.14268, 2022.
- Probabilistic (logic) programming concepts. Machine Learning, 2015.
- A knowledge compilation map. Journal of Artificial Intelligence Research, 17:229–264, 2002.
- A compositional atlas of tractable circuit operations for probabilistic inference. Advances in Neural Information Processing Systems, 34, 2021.
- Approximate inference for neural probabilistic logic programming. In KR, 2021b.
- Scallop: From probabilistic deductive databases to scalable differentiable reasoning. NeurIPS, 2021.
- DeepStochLog: Neural Stochastic Logic Programming. In AAAI, 2022.
- A-nesi: A scalable approximate method for probabilistic neurosymbolic inference. arXiv preprint arXiv:2212.12393, 2022.
- Toward causal representation learning. Proceedings of the IEEE, 2021.
- Robustly disentangled causal mechanisms: Validating deep representations for interventional robustness. In International Conference on Machine Learning, pages 6056–6065. PMLR, 2019.
- Self-supervised learning with data augmentations provably isolates content from style. Advances in neural information processing systems, 34:16451–16467, 2021.
- Dataset shift in machine learning. Mit Press, 2008.
- Glancenets: Interpretabile, leak-proof concept-based models. NeurIPS, 2022.
- Autoencoders, minimum description length and helmholtz free energy. Advances in neural information processing systems, 6, 1993.
- Stochastic backpropagation and approximate inference in deep generative models. In International conference on machine learning, 2014.
- Auto-encoding variational bayes. In International conference on machine learning, 2014.
- From variational to deterministic autoencoders. In ICLR, 2020.
- VAEL: Bridging Variational Autoencoders and Probabilistic Logic Programming. NeurIPS, 2022.
- Challenging common assumptions in the unsupervised learning of disentangled representations. In ICML, 2019.
- Weakly-supervised disentanglement without compromises. In International Conference on Machine Learning, pages 6348–6359. PMLR, 2020a.
- Weakly supervised disentanglement with guarantees. In ICLR, 2019.
- When does label smoothing help? Advances in neural information processing systems, 32, 2019.
- Energy-based models for continual learning. In Energy Based Models Workshop-ICLR 2021, 2021.
- Closed loop neural-symbolic learning via integrating neural perception, grammar parsing, and symbolic reasoning. In International Conference on Machine Learning, pages 5884–5894. PMLR, 2020.
- Mitigating neural network overconfidence with logit normalization. In International Conference on Machine Learning, pages 23631–23644. PMLR, 2022.
- Calibrating deep neural networks using focal loss. Advances in Neural Information Processing Systems, 33:15288–15299, 2020.
- On mixup regularization. The Journal of Machine Learning Research, 23(1):14632–14662, 2022.
- Explainable object-induced action decision for autonomous vehicles. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
- Concept bottleneck model with additional unsupervised concepts. IEEE Access, 10:41758–41765, 2022.
- Shortcut learning in deep neural networks. Nature Machine Intelligence, 2(11):665–673, 2020.
- Unmasking clever hans predictors and assessing what machines really learn. Nature communications, 10(1):1–8, 2019.
- Noise or signal: The role of image backgrounds in object recognition. In ICLR, 2020.
- Making deep neural networks right for the right scientific reasons by interacting with their explanations. Nature Machine Intelligence, 2(8):476–486, 2020.
- Right for the right reasons: training differentiable models by constraining their explanations. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, pages 2662–2670, 2017.
- Learning explanations that are hard to vary. In International Conference on Learning Representations, 2020.
- Leveraging explanations in interactive machine learning: An overview. Frontiers in Artificial Intelligence, 2023.
- Interactive disentanglement: Learning concepts by interacting with their prototype representations. arXiv preprint arXiv:2112.02290, 2021.
- Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8):1798–1828, 2013.
- Nonlinear independent component analysis for principled disentanglement in unsupervised deep learning. Patterns, 4(10), 2023.
- Variational autoencoders and nonlinear ICA: A unifying framework. In AISTATS, 2020.
- Function classes for identifiable nonlinear independent component analysis. Advances in Neural Information Processing Systems, 35:16946–16961, 2022.
- Independent mechanism analysis, a new concept? Advances in neural information processing systems, 34:28233–28248, 2021.
- Interventional causal representation learning. In International conference on machine learning, pages 372–407. PMLR, 2023.
- Causal component analysis. arXiv preprint arXiv:2305.17225, 2023.
- Nonparametric identifiability of causal representations from unknown interventions. arXiv preprint arXiv:2306.00542, 2023.
- On the relationship between disentanglement and multi-task learning. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 625–641. Springer, 2022.
- Synergies between disentanglement and sparsity: generalization and identifiability in multi-task learning. In International Conference on Machine Learning, pages 18171–18206. PMLR, 2023.
- Leveraging sparse and shared feature activations for disentangled representation learning. arXiv preprint arXiv:2304.07939, 2023.
- Disentangling factors of variations using few labels. In International Conference on Learning Representations, 2020b.
- Semi-supervised stylegan for disentanglement learning. In Proceedings of the 37th International Conference on Machine Learning, pages 7360–7369, 2020.
- From perception to programs: regularize, overparameterize, and amortize. In International Conference on Machine Learning, pages 33616–33631. PMLR, 2023.
- Deep symbolic learning: Discovering symbols and rules from perceptions. arXiv preprint arXiv:2208.11561, 2022.
- Concept bottleneck models. In International Conference on Machine Learning, pages 5338–5348. PMLR, 2020.
- Concept embedding models. arXiv preprint arXiv:2209.09056, 2022.
- On a convex logic fragment for learning and reasoning. IEEE Transactions on Fuzzy Systems, 2018.
- Thomas M Cover. Elements of information theory. John Wiley & Sons, 1999.
- A pseudo-semantic loss for deep generative models with logical constraints. In Knowledge and Logical Reasoning in the Era of Data-driven Learning Workshop, July 2023.
- Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
- Tommaso Carraro. LTNtorch: PyTorch implementation of Logic Tensor Networks, mar 2022. URL https://doi.org/10.5281/zenodo.6394282.
- Adam: A method for stochastic optimization. In Yoshua Bengio and Yann LeCun, editors, 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. URL http://arxiv.org/abs/1412.6980.
- Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 28, 2015.