- The paper introduces GLGExplainer, which transforms local GNN explanations into global Boolean logic rules.
- It clusters local subgraph interpretations into prototypical concepts to reveal high-level decision patterns.
- Evaluations on synthetic and real datasets demonstrate high fidelity in capturing GNN behavior and identifying key features like nitro groups.
Global Explainability of GNNs via Logic Combination of Learned Concepts
The research presented in the paper addresses a significant challenge in graph neural networks (GNNs): providing global explanations for model behavior. While local explanation methods have been extensively developed, they often fail to encapsulate the global patterns and interactions that a GNN learns from its input data. This paper introduces GLGExplainer, a novel approach that constructs global explanations for GNNs using Boolean combinations of learned graphical concepts.
GLGExplainer operates by first extracting local explanations from a GNN's predictions using existing local explanation methods, like PGExplainer. These local explanations, which are typically subgraphs of the input data, are embedded into a fixed-size representation. The embeddings serve as a basis for identifying prototypical examples, which are then grouped into clusters representing high-level concepts. These clusters are crucial as they provide interpretable units for global explanations.
A unique aspect of GLGExplainer is its use of a differentiable architecture that combines these concepts into logic formulas. These formulas aim to mimic the GNN's decision-making process by logically merging the identified concepts. The approach leverages entropy-based logic explained networks (E-LENs) to distill these explanations into human-understandable rules expressed in Boolean logic. This setup enables GLGExplainer not only to offer insights into the model's decision structures but also to identify and potentially correct incorrect rules learned by the model.
The experimental evaluation of GLGExplainer includes synthetic and real-world datasets, such as BAMultiShapes, Mutagenicity, and the Hospital Interaction Network (HIN). Notably, GLGExplainer demonstrated high fidelity and accuracy in capturing the GNNs' behavior across these datasets. For instance, it successfully identified common mutagenicity indicators such as nitro groups in the Mutagenicity dataset. On BAMultiShapes, GLGExplainer effectively reconstructed the underlying logic of the dataset, albeit highlighting discrepancies between the model and ground-truth labeling.
Overall, GLGExplainer marks a significant step towards enhancing the interpretability of GNNs through global explanations. By transforming local explanations into coherent global insights, it offers substantial utility for debugging and better understanding complex models. Future research in this area could focus on extending GLGExplainer to various types of neural networks and exploring its applications in domains where understanding model logic is imperative for deployment. Additionally, improvements might include refining the concept clustering mechanism or expanding the approach to handle dynamic or evolving graphs. Such advances would further align model predictions with human domains of knowledge, enhancing trust and accountability in AI systems.