Interpretable Neural-Symbolic Concept Reasoning (2304.14068v2)
Abstract: Deep learning methods are highly accurate, yet their opaque decision process prevents them from earning full human trust. Concept-based models aim to address this issue by learning tasks based on a set of human-understandable concepts. However, state-of-the-art concept-based models rely on high-dimensional concept embedding representations which lack a clear semantic meaning, thus questioning the interpretability of their decision process. To overcome this limitation, we propose the Deep Concept Reasoner (DCR), the first interpretable concept-based model that builds upon concept embeddings. In DCR, neural networks do not make task predictions directly, but they build syntactic rule structures using concept embeddings. DCR then executes these rules on meaningful concept truth degrees to provide a final interpretable and semantically-consistent prediction in a differentiable manner. Our experiments show that DCR: (i) improves up to +25% w.r.t. state-of-the-art interpretable concept-based models on challenging benchmarks (ii) discovers meaningful logic rules matching known ground truths even in the absence of concept supervision during training, and (iii), facilitates the generation of counterfactual examples providing the learnt rules as guidance.
- Meaningfully explaining model mistakes using conceptual counterfactuals. arXiv preprint arXiv:2106.12723, 2021.
- ACGIH®. American conference of governmental industrial hygienists: Tlvs and beis based on the documentation of the threshold limit values for chemical substances and physical agents and biological exposure indices. American Conference of Governmental Industrial Hygienists Washington, DC, USA, 2016.
- Sanity checks for saliency maps. Advances in neural information processing systems, 31, 2018.
- Embed2sym-scalable neuro-symbolic reasoning via clustered embeddings. In Proceedings of the International Conference on Principles of Knowledge Representation and Reasoning, volume 19, pp. 421–431, 2022.
- Global explainability of gnns via logic combination of learned concepts. arXiv preprint arXiv:2210.07147, 2022.
- Logic tensor networks. Artificial Intelligence, 303:103649, 2022.
- Entropy-based logic explanations of neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 6046–6054, 2022.
- Categorical foundations of explainable ai: A unifying formalism of structures and semantics. arXiv preprint arXiv:2304.14094, 2023.
- Breiman, L. Classification and regression trees. Routledge, 2017.
- The role of explanations on trust and reliance in clinical decision support systems. In 2015 international conference on healthcare informatics, pp. 160–169. IEEE, 2015.
- XGBoost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, pp. 785–794, 2016.
- Concept whitening for interpretable image recognition. Nature Machine Intelligence, 2(12):772–782, 2020.
- Logic explained networks. Artificial Intelligence, 314:103822, 2023.
- Neural logic machines. arXiv preprint arXiv:1904.11694, 2019.
- Who is afraid of black box algorithms? On the epistemological and ethical basis of trust in medical AI. Journal of Medical Ethics, 47(5):329–335, 2021.
- Concept embedding models. Advances in Neural Information Processing Systems, 35, 2022.
- Towards robust metrics for concept representation evaluation. AAAI, 2023.
- EUGDPR. GDPR. General data protection regulation, 2017.
- Fast graph representation learning with pytorch geometric. arXiv preprint arXiv:1903.02428, 2019.
- Forgy, E. W. Cluster analysis of multivariate data: efficiency versus interpretability of classifications. biometrics, 21:768–769, 1965.
- Algorithmic concept-based explainable reasoning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 6685–6693, 2022.
- Interpretation of neural networks is fragile. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 3681–3688, 2019a.
- Towards automatic concept-based explanations. arXiv preprint arXiv:1902.03129, 2019b.
- Hájek, P. Metamathematics of fuzzy logic, volume 4. 2013.
- A simple generalisation of the area under the roc curve for multiple class classification problems. Machine learning, 45(2):171–186, 2001.
- The role of polyphenols in terrestrial ecosystem nutrient cycling. Trends in ecology & evolution, 15(6):238–243, 2000.
- Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.
- Learning by abstraction: The neural state machine. Advances in Neural Information Processing Systems, 32, 2019.
- Now you see me (cme): concept-based model extraction. arXiv preprint arXiv:2010.13233, 2020.
- Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In International conference on machine learning, pp. 2668–2677. PMLR, 2018.
- The (un) reliability of saliency methods. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning, pp. 267–280. Springer, 2019.
- Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Concept bottleneck models. In International Conference on Machine Learning, pp. 5338–5348. PMLR, 2020.
- Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision, pp. 3730–3738, 2015.
- Lo Piano, S. Ethical principles in machine learning and artificial intelligence: cases from the field and possible ways forward. Humanities and Social Sciences Communications, 7(1):1–7, 2020.
- Gcexplainer: Human-in-the-loop concept-based explanations for graph neural networks. arXiv preprint arXiv:2107.11889, 2021.
- Encoding concepts in graph neural networks. arXiv preprint arXiv:2207.13586, 2022.
- Promises and pitfalls of black-box concept learning models. arXiv preprint arXiv:2106.13314, 2021.
- Deepproblog: Neural probabilistic logic programming. Advances in Neural Information Processing Systems, 31, 2018.
- The neuro-symbolic concept learner: Interpreting scenes, words, and sentences from natural supervision. arXiv preprint arXiv:1904.12584, 2019.
- Relational neural machines. arXiv preprint arXiv:2002.02193, 2020a.
- Integrating learning and reasoning with deep logic models. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2019, Würzburg, Germany, September 16–20, 2019, Proceedings, Part II, pp. 517–532. Springer, 2020b.
- Lyrics: A general interface layer to integrate logic inference and deep learning. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 283–298. Springer, 2020c.
- Miller, G. A. The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychological review, 63(2):81, 1956.
- Perceptrons: An introduction to computational geometry. MIT press, 1969.
- Weisfeiler and leman go neural: Higher-order graph neural networks. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 4602–4609, 2019.
- Tudataset: A collection of benchmark datasets for learning with graphs. arXiv preprint arXiv:2007.08663, 2020.
- Pytorch: An imperative style, high-performance deep learning library. arXiv preprint arXiv:1912.01703, 2019.
- Scikit-learn: Machine learning in python. the Journal of machine Learning research, 12:2825–2830, 2011.
- ” why should i trust you?” explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp. 1135–1144, 2016.
- Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5):206–215, 2019.
- Synthesis of 4h-chromene, coumarin, 12h-chromeno [2, 3-d] pyrimidine derivatives and some of their antimicrobial and cytotoxicity activities. European journal of medicinal chemistry, 46(2):765–772, 2011.
- The graph neural network model. IEEE transactions on neural networks, 20(1):61–80, 2008.
- Shen, M. W. Trust in AI: Interpretability is not necessary or sufficient, while black-box interaction is necessary and sufficient. arXiv preprint arXiv:2202.05302, 2022.
- Verhulst, P. F. Resherches mathematiques sur la loi d’accroissement de la population. Nouveaux memoires de l’academie royale des sciences, 18:1–41, 1845.
- Counterfactual explanations without opening the black box: Automated decisions and the gdpr. Harv. JL & Tech., 31:841, 2017.
- Neural-symbolic integration for interactive learning and conceptual grounding. arXiv preprint arXiv:2112.11805, 2021.
- The caltech-ucsd birds-200-2011 dataset, 2011.
- Deepstochlog: Neural stochastic logic programming. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 10090–10100, 2022.
- Global concept-based interpretability for graph neural networks via neuron analysis. arXiv preprint arXiv:2208.10609, 2022.
- Neurasp: Embracing neural networks into answer set programming. In 29th International Joint Conference on Artificial Intelligence (IJCAI 2020), 2020.
- On the (in) fidelity and sensitivity of explanations. Advances in Neural Information Processing Systems, 32, 2019.
- On completeness-aware concept-based explanations in deep neural networks. Advances in Neural Information Processing Systems, 33:20554–20565, 2020.
- Gnnexplainer: Generating explanations for graph neural networks. Advances in neural information processing systems, 32, 2019.