Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Structure Your Data: Towards Semantic Graph Counterfactuals (2403.06514v2)

Published 11 Mar 2024 in cs.CV and cs.AI

Abstract: Counterfactual explanations (CEs) based on concepts are explanations that consider alternative scenarios to understand which high-level semantic features contributed to particular model predictions. In this work, we propose CEs based on the semantic graphs accompanying input data to achieve more descriptive, accurate, and human-aligned explanations. Building upon state-of-the-art (SoTA) conceptual attempts, we adopt a model-agnostic edit-based approach and introduce leveraging GNNs for efficient Graph Edit Distance (GED) computation. With a focus on the visual domain, we represent images as scene graphs and obtain their GNN embeddings to bypass solving the NP-hard graph similarity problem for all input pairs, an integral part of the CE computation process. We apply our method to benchmark and real-world datasets with varying difficulty and availability of semantic annotations. Testing on diverse classifiers, we find that our CEs outperform previous SoTA explanation models based on semantics, including both white and black-box as well as conceptual and pixel-level approaches. Their superiority is proven quantitatively and qualitatively, as validated by human subjects, highlighting the significance of leveraging semantic edges in the presence of intricate relationships. Our model-agnostic graph-based approach is widely applicable and easily extensible, producing actionable explanations across different contexts.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Meaningfully debugging model mistakes using conceptual counterfactual explanations. In International Conference on Machine Learning, pp.  66–88. PMLR, 2022.
  2. Diffusion visual counterfactual explanations. arXiv preprint arXiv:2210.11841, 2022.
  3. Unsupervised inductive graph-level representation learning via graph-graph proximity. arXiv preprint arXiv:1904.01098, 2019.
  4. Robust counterfactual explanations on graph neural networks. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems, volume 34, pp.  5644–5655. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper/2021/file/2c8c3a57383c63caef6724343eb62257-Paper.pdf.
  5. Semantics and explanation: why counterfactual explanations produce adversarial examples in deep neural networks. arXiv preprint arXiv:2012.10076, 2020.
  6. Reltr: Relation transformer for scene graph generation, 2023.
  7. Deccan Chronicle. Cyclists without helmet, 2016. URL https://www.deccanchronicle.com/nation/current-affairs/030316/motorists-scramble-for-driving-licence.html. [Online; accessed February 1, 2024].
  8. Choose your data wisely: A framework for semantic counterfactuals. In Elkind, E. (ed.), Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-23, pp.  382–390. International Joint Conferences on Artificial Intelligence Organization, 8 2023. doi: 10.24963/ijcai.2023/43. URL https://doi.org/10.24963/ijcai.2023/43. Main Track.
  9. Speeding up graph edit distance computation through fast bipartite matching. In International Workshop on Graph-Based Representations in Pattern Recognition, pp.  102–111. Springer, 2011.
  10. Fast graph representation learning with PyTorch Geometric. In ICLR Workshop on Representation Learning on Graphs and Manifolds, 2019.
  11. Conceptual edits as counterfactual explanations. In AAAI Spring Symposium: MAKE, 2022.
  12. Counterfactual visual explanations. In Proceedings of the 36th International Conference on Machine Learning, pp.  2376–2384, 2019.
  13. The pyramid match kernel: Efficient learning with sets of features. Journal of Machine Learning Research, 8(4), 2007.
  14. Deep residual learning for image recognition, 2015.
  15. Gpt-4v exhibits human-like performance in biomedical image classification. bioRxiv, pp.  2023–12, 2024.
  16. Gqa: A new dataset for real-world visual reasoning and compositional question answering. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  6700–6709, 2019.
  17. Action genome: Actions as compositions of spatio-temporal scene graphs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  10236–10247, 2020.
  18. A shortest augmenting path algorithm for dense and sparse linear assignment problems. Computing, 38(4):325–340, 1987.
  19. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
  20. Visual genome: Connecting language and vision using crowdsourced dense image annotations. International journal of computer vision, 123(1):32–73, 2017.
  21. Biobert: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics, 36(4):1234–1240, 2020.
  22. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation, 2022. URL https://arxiv.org/abs/2201.12086.
  23. Graph matching networks for learning the similarity of graph structured objects. In International conference on machine learning, pp.  3835–3845. PMLR, 2019.
  24. Cf-gnnexplainer: Counterfactual explanations for graph neural networks. In International Conference on Artificial Intelligence and Statistics, pp.  4499–4511. PMLR, 2022.
  25. Miller, G. A. Wordnet: a lexical database for english. Communications of the ACM, 38(11):39–41, 1995.
  26. Matching node embeddings for graph similarity. In Thirty-first AAAI conference on artificial intelligence, 2017.
  27. Glove: Global vectors for word representation. In Empirical Methods in Natural Language Processing (EMNLP), pp.  1532–1543, 2014. URL http://www.aclweb.org/anthology/D14-1162.
  28. Face: feasible and actionable counterfactual explanations. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, pp.  344–350, 2020.
  29. Greed: A neural framework for learning graph distance functions. In Advances in Neural Information Processing Systems, 2022.
  30. ” why should i trust you?” explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp.  1135–1144, 2016.
  31. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5):206–215, 2019.
  32. A distance measure between attributed relational graphs for pattern recognition. IEEE transactions on systems, man, and cybernetics, (3):353–362, 1983.
  33. Grakel: A graph kernel library in python. J. Mach. Learn. Res., 21(54):1–5, 2020.
  34. Counterfactual explanations can be manipulated. Advances in neural information processing systems, 34:62–75, 2021.
  35. Gemini: a family of highly capable multimodal models. arXiv preprint arXiv:2312.11805, 2023.
  36. Making heads or tails: Towards semantically consistent visual counterfactuals. arXiv preprint arXiv:2203.12892, 2022.
  37. Graph attention networks. arXiv preprint arXiv:1710.10903, 2017.
  38. Counterfactual explanations without opening the black box: Automated decisions and the gdpr. Harv. JL & Tech., 31:841, 2017.
  39. The caltech-ucsd birds-200-2011 dataset. 2011.
  40. Deep graph library: A graph-centric, highly-performant package for graph neural networks. arXiv preprint arXiv:1909.01315, 2019.
  41. Williams, C. On a connection between kernel pca and metric multidimensional scaling. Advances in neural information processing systems, 13, 2000.
  42. Can gpt-4v (ision) serve medical applications? case studies on gpt-4v for multimodal medical diagnosis. arXiv preprint arXiv:2310.09909, 2023.
  43. Unified visual-semantic embeddings: Bridging vision and language with structured meaning representations. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  6602–6611, 2019. URL https://api.semanticscholar.org/CorpusID:119284150.
  44. How powerful are graph neural networks? arXiv preprint arXiv:1810.00826, 2018.
  45. Gnnexplainer: Generating explanations for graph neural networks. Advances in neural information processing systems, 32, 2019.
  46. The smarty4covid dataset and knowledge base: a framework enabling interpretable analysis of audio signals. arXiv preprint arXiv:2307.05096, 2023.
  47. Places: A 10 million image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017.
Citations (2)

Summary

We haven't generated a summary for this paper yet.