Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Accurate and Scalable Estimation of Epistemic Uncertainty for Graph Neural Networks (2401.03350v2)

Published 7 Jan 2024 in cs.LG and stat.ML

Abstract: While graph neural networks (GNNs) are widely used for node and graph representation learning tasks, the reliability of GNN uncertainty estimates under distribution shifts remains relatively under-explored. Indeed, while post-hoc calibration strategies can be used to improve in-distribution calibration, they need not also improve calibration under distribution shift. However, techniques which produce GNNs with better intrinsic uncertainty estimates are particularly valuable, as they can always be combined with post-hoc strategies later. Therefore, in this work, we propose G-$\Delta$UQ, a novel training framework designed to improve intrinsic GNN uncertainty estimates. Our framework adapts the principle of stochastic data centering to graph data through novel graph anchoring strategies, and is able to support partially stochastic GNNs. While, the prevalent wisdom is that fully stochastic networks are necessary to obtain reliable estimates, we find that the functional diversity induced by our anchoring strategies when sampling hypotheses renders this unnecessary and allows us to support G-$\Delta$UQ on pretrained models. Indeed, through extensive evaluation under covariate, concept and graph size shifts, we show that G-$\Delta$UQ leads to better calibrated GNNs for node and graph classification. Further, it also improves performance on the uncertainty-based tasks of out-of-distribution detection and generalization gap estimation. Overall, our work provides insights into uncertainty estimation for GNNs, and demonstrates the utility of G-$\Delta$UQ in obtaining reliable estimates.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (61)
  1. On the bottleneck of graph neural networks and its practical implications. In Proc. Int. Conf. on Learning Representations (ICLR), 2021.
  2. Out of distribution detection via neural network anchoring. In Asian Conference on Machine Learning, ACML 2022, 12-14 December 2022, Hyderabad, India, 2022.
  3. Size-invariant graph representations for graph classification extrapolations. In Proc. Int. Conf. on Machine Learning (ICML), 2021.
  4. Weight uncertainty in neural network. In Proc. Int. Conf. on Machine Learning (ICML), 2015.
  5. Sizeshiftreg: a regularization method for improving size-generalization in graph neural networks. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2022.
  6. Principal neighbourhood aggregation for graph nets. In NeurIPS, 2020.
  7. Torchmetrics - measuring reproducibility in pytorch, 2022. URL https://github.com/Lightning-AI/torchmetrics.
  8. A closer look at distribution shifts and out-of-distribution generalization on graphs. In NeurIPS 2021 Workshop on Distribution Shifts: Connecting Methods and Applications, 2021.
  9. Benchmarking graph neural networks. CoRR, 2020.
  10. Graph neural networks with learnable structural and positional representations. In Proc. Int. Conf. on Learning Representations (ICLR), 2022a.
  11. Long range graph benchmark. In Proc. Adv. in Neural Information Processing Systems NeurIPS, Datasets and Benchmark Track, 2022b.
  12. Leveraging unlabeled data to predict out-of-distribution performance. In Proc. Int. Conf. on Learning Representations (ICLR), 2022.
  13. Utilising graph machine learning within drug discovery and development. CoRR, abs/2012.05716, 2020.
  14. GOOD: A graph out-of-distribution benchmark. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), Benchmark Track, 2022.
  15. Predicting with confidence on unseen distributions. In ICCV, 2021.
  16. On calibration of modern neural networks. In Proc. of the Int. Conf. on Machine Learning, (ICML), 2017.
  17. Calibration of neural networks using splines. In Proc. Int. Conf. on Learning Representations (ICLR), 2021.
  18. Bayesian graph neural networks with adaptive connection sampling. In ICML, 2020.
  19. A baseline for detecting misclassified and out-of-distribution examples in neural networks. In Proc. Int. Conf. on Learning Representations (ICLR), 2017.
  20. Deep anomaly detection with outlier exposure. In Proc. Int. Conf. on Learning Representations (ICLR), 2019.
  21. Unsolved problems in ML safety. CoRR, abs/2109.13916, 2021.
  22. Scaling out-of-distribution detection for real-world settings. In Proc. Int. Conf. on Machine Learning (ICML), 2022a.
  23. Pixmix: Dreamlike pictures comprehensively improve safety measures. In Proc. Int. Conf. on Computer Vision and Pattern Recognition (CVPR), 2022b.
  24. Probabilistic backpropagation for scalable learning of bayesian neural networks. In Proc. Int. Conf. on Machine Learning (ICML), 2015.
  25. What makes graph neural networks miscalibrated? In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2022.
  26. Aleatoric and epistemic uncertainty in machine learning: an introduction to concepts and methods. Mach. Learn., 110(3), 2021.
  27. Predicting the generalization gap in deep networks with margin distributions. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
  28. Semi-supervised classification with graph convolutional networks. In ICLR, 2017.
  29. Pytorch-ood: A library for out-of-distribution detection based on pytorch. In Workshop at the Proc. Int. Conf. on Computer Vision and Pattern Recognition CVPR, 2022.
  30. Understanding attention and generalization in graph neural networks. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2019.
  31. Improving model calibration with accuracy versus uncertainty optimization. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2020.
  32. Beyond temperature scaling: Obtaining well-calibrated multi-class probabilities with dirichlet calibration. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2019.
  33. Verified uncertainty calibration. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2019.
  34. Simple and scalable predictive uncertainty estimation using deep ensembles. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2017.
  35. A simple unified framework for detecting out-of-distribution samples and adversarial attacks. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2018.
  36. Energy-based out-of-distribution detection. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2020.
  37. Revisiting the calibration of modern neural networks. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2021.
  38. Tudataset: A collection of benchmark datasets for learning with graphs. In ICML 2020 Workshop on Graph Representation Learning and Beyond (GRL+ 2020), 2020. URL www.graphlearning.io.
  39. Attending to graph transformers. CoRR, abs/2302.04181, 2023.
  40. Obtaining well calibrated probabilities using bayesian binning. In Proc. Conf. on Adv. of Artificial Intelligence (AAAI), 2015.
  41. Learning to extrapolate: A transductive approach. In Proc. Int. Conf. on Learning Representations (ICLR), 2023.
  42. Predicting out-of-domain generalization with local manifold smoothness. CoRR, abs/2207.02093, 2022.
  43. Can you trust your model’s uncertainty? evaluating predictive uncertainty under dataset shift. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2019.
  44. Intra order-preserving functions for calibration of multi-class neural networks. Advances in Neural Information Processing Systems, 33:13456–13467, 2020.
  45. Do bayesian neural networks need to be fully stochastic? In AISTATS, 2023.
  46. Single model uncertainty estimation via stochastic data centering. In Proc. Adv. in Neural Information Processing Systems (NeurIPS), 2022.
  47. Understanding over-squashing and bottlenecks on graphs via curvature. In Proc. Int. Conf. on Learning Representations ICLR, 2022.
  48. A closer look at scoring functions and generalization prediction. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.  1–5. IEEE, 2023a.
  49. A closer look at model adaptation using feature distortion and simplicity bias. In Proc. Int. Conf. on Learning Representations (ICLR), 2023b.
  50. Graph attention networks. In ICLR, 2018.
  51. Vim: Out-of-distribution with virtual-logit matching. In Proc. Int. Conf. on Computer Vision and Pattern Recognition (CVPR), 2022a.
  52. Equivariant and stable positional encoding for more powerful graph neural networks. In Proc. Int. Conf. on Learning Representations (ICLR), 2022b.
  53. Be confident! towards trustworthy graph neural networks via confidence calibration. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2021.
  54. A Fine-Grained Analysis on Distribution Shift. In Proc. Int. Conf. on Learning Representations (ICLR), 2022.
  55. Bayesian deep learning and a probabilistic perspective of generalization. In Proc. Adv. in Neural Information Processing Systems NeurIPS, 2020.
  56. How powerful are graph neural networks? In ICLR, 2019.
  57. From local structures to size generalization in graph neural networks. In International Conference on Machine Learning, pp.  11975–11986. PMLR, 2021.
  58. Transforming classifier scores into accurate multiclass probability estimates. In Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining, pp.  694–699, 2002.
  59. Mix-n-match : Ensemble and compositional methods for uncertainty calibration in deep learning. In Proc. Int. Conf. on Machine Learning (ICML), 2020.
  60. Bayesian graph convolutional neural networks for semi-supervised classification. In AAAI, 2019.
  61. A survey on deep graph generation: Methods and applications. In Learning on Graphs Conference (LoG), 2022.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets