Accurate and Scalable Estimation of Epistemic Uncertainty for Graph Neural Networks (2309.10976v1)
Abstract: Safe deployment of graph neural networks (GNNs) under distribution shift requires models to provide accurate confidence indicators (CI). However, while it is well-known in computer vision that CI quality diminishes under distribution shift, this behavior remains understudied for GNNs. Hence, we begin with a case study on CI calibration under controlled structural and feature distribution shifts and demonstrate that increased expressivity or model size do not always lead to improved CI performance. Consequently, we instead advocate for the use of epistemic uncertainty quantification (UQ) methods to modulate CIs. To this end, we propose G-$\Delta$UQ, a new single model UQ method that extends the recently proposed stochastic centering framework to support structured data and partial stochasticity. Evaluated across covariate, concept, and graph size shifts, G-$\Delta$UQ not only outperforms several popular UQ methods in obtaining calibrated CIs, but also outperforms alternatives when CIs are used for generalization gap prediction or OOD detection. Overall, our work not only introduces a new, flexible GNN UQ method, but also provides novel insights into GNN CIs on safety-critical tasks.
- SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell., 2012.
- On the bottleneck of graph neural networks and its practical implications. In Proc. Int. Conf. on Learning Representations (ICLR), 2021.
- Size-invariant graph representations for graph classification extrapolations. In Proc. Int. Conf. on Machine Learning (ICML), 2021.
- Weight uncertainty in neural network. In Proc. Int. Conf. on Machine Learning (ICML), 2015.
- Sizeshiftreg: a regularization method for improving size-generalization in graph neural networks. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2022.
- Learning causally invariant representations for out-of-distribution generalization on graphs. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2022.
- Tree mover’s distance: Bridging graph metrics and stability of graph neural networks. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2022.
- Principal neighbourhood aggregation for graph nets. In NeurIPS, 2020.
- Torchmetrics - measuring reproducibility in pytorch, 2022. URL https://github.com/Lightning-AI/torchmetrics.
- A closer look at distribution shifts and out-of-distribution generalization on graphs. In NeurIPS 2021 Workshop on Distribution Shifts: Connecting Methods and Applications, 2021.
- Benchmarking graph neural networks. CoRR, 2020.
- Graph neural networks with learnable structural and positional representations. In Proc. Int. Conf. on Learning Representations (ICLR), 2022a.
- Long range graph benchmark. In Proc. Adv. in Neural Information Processing Systems NeurIPS, Datasets and Benchmark Track, 2022b.
- Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In Proc. Int. Conf. on Machine Learning (ICML), 2016.
- Leveraging unlabeled data to predict out-of-distribution performance. In Proc. Int. Conf. on Learning Representations (ICLR), 2022.
- Utilising graph machine learning within drug discovery and development. CoRR, abs/2012.05716, 2020.
- GOOD: A graph out-of-distribution benchmark. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), Benchmark Track, 2022.
- Predicting with confidence on unseen distributions. In ICCV, 2021.
- On calibration of modern neural networks. In Proc. of the Int. Conf. on Machine Learning, (ICML), 2017.
- A generalization of vit/mlp-mixer to graphs. CoRR, abs/2212.13350, 2022.
- A baseline for detecting misclassified and out-of-distribution examples in neural networks. In Proc. Int. Conf. on Learning Representations (ICLR), 2017.
- Deep anomaly detection with outlier exposure. In Proc. Int. Conf. on Learning Representations (ICLR), 2019.
- Unsolved problems in ML safety. CoRR, abs/2109.13916, 2021.
- Scaling out-of-distribution detection for real-world settings. In Proc. Int. Conf. on Machine Learning (ICML), 2022a.
- Pixmix: Dreamlike pictures comprehensively improve safety measures. In Proc. Int. Conf. on Computer Vision and Pattern Recognition (CVPR), 2022b.
- What makes graph neural networks miscalibrated? In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2022.
- Predicting the generalization gap in deep networks with margin distributions. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
- Jurygcn: Quantifying jackknife uncertainty on graph convolutional networks. In Proc. Int. Conf. on Knowledge Discovery & Data Mining, KDD, 2022.
- Semi-supervised classification with graph convolutional networks. In ICLR, 2017.
- Pytorch-ood: A library for out-of-distribution detection based on pytorch. In Workshop at the Proc. Int. Conf. on Computer Vision and Pattern Recognition CVPR, 2022.
- Understanding attention and generalization in graph neural networks. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2019.
- Verified uncertainty calibration. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2019.
- Simple and scalable predictive uncertainty estimation using deep ensembles. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2017.
- A simple unified framework for detecting out-of-distribution samples and adversarial attacks. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2018.
- Distance encoding: Design provably more powerful neural networks for graph representation learning. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2020.
- Energy-based out-of-distribution detection. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2020.
- Revisiting the calibration of modern neural networks. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2021.
- Tudataset: A collection of benchmark datasets for learning with graphs. In ICML 2020 Workshop on Graph Representation Learning and Beyond (GRL+ 2020), 2020. URL www.graphlearning.io.
- Attending to graph transformers. CoRR, abs/2302.04181, 2023.
- Obtaining well calibrated probabilities using bayesian binning. In Proc. Conf. on Adv. of Artificial Intelligence (AAAI), 2015.
- Learning to extrapolate: A transductive approach. In Proc. Int. Conf. on Learning Representations (ICLR), 2023.
- Predicting out-of-domain generalization with local manifold smoothness. CoRR, abs/2207.02093, 2022.
- Can you trust your model’s uncertainty? evaluating predictive uncertainty under dataset shift. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2019.
- Recipe for a General, Powerful, Scalable Graph Transformer. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2022.
- Do bayesian neural networks need to be fully stochastic? In AISTATS, 2023.
- Single model uncertainty estimation via stochastic data centering. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2022.
- Understanding over-squashing and bottlenecks on graphs via curvature. In Proc. Int. Conf. on Learning Representations ICLR, 2022.
- A closer look at scoring functions and generalization prediction. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1–5. IEEE, 2023a.
- A closer look at model adaptation using feature distortion and simplicity bias. In Proc. Int. Conf. on Learning Representations (ICLR), 2023b.
- Graph attention networks. In ICLR, 2018.
- Vim: Out-of-distribution with virtual-logit matching. In Proc. Int. Conf. on Computer Vision and Pattern Recognition (CVPR), 2022a.
- Equivariant and stable positional encoding for more powerful graph neural networks. In Proc. Int. Conf. on Learning Representations (ICLR), 2022b.
- Be confident! towards trustworthy graph neural networks via confidence calibration. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2021.
- A Fine-Grained Analysis on Distribution Shift. In Proc. Int. Conf. on Learning Representations (ICLR), 2022.
- How powerful are graph neural networks? In ICLR, 2019.
- Optimization of graph neural networks: Implicit acceleration by skip connections and more depth. In Proc. Int. Conf. on Machine Learning (ICML), 2021.
- Groupinn: Grouping-based interpretable neural network for classification of limited, noisy brain data. In Proc. Int. Conf. on Knowledge Discovery & Data Mining, KDD, 2019.
- Graph R-CNN for scene graph generation. In Proc. Euro. Conf. on Computer Vision (ECCV), 2018.
- From local structures to size generalization in graph neural networks. In International Conference on Machine Learning, pp. 11975–11986. PMLR, 2021.
- Link prediction based on graph neural networks. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2018.
- From stars to subgraphs: Uplifting any GNN with local structure awareness. In Proc. Int. Conf. on Learning Representations (ICLR), 2022.
- A survey on deep graph generation: Methods and applications. In Learning on Graphs Conference (LoG), 2022.