Papers
Topics
Authors
Recent
Search
2000 character limit reached

Bayesian Graph Neural Networks

Updated 26 February 2026
  • Bayesian Graph Neural Networks are graph-based learning frameworks that explicitly model uncertainty by incorporating probabilistic priors over graph structure, parameters, and latent variables.
  • They employ methodologies like variational inference, Monte Carlo sampling, and Beta-process priors to enable robust predictions in noisy, sparse, or partially observed graph data.
  • These models are applied in tasks such as node classification, link prediction, and molecular property regression, demonstrating improved calibration and resilience to overconfidence.

A Bayesian Graph Neural Network (BGNN) is any graph neural network architecture that replaces deterministic or ad hoc point estimates by explicit probability modeling—either over the graph structure, the network parameters, or higher-order latent variables—to characterize uncertainty in inference and prediction. Across the literature, BGNNs are instantiated using diverse frameworks: non-parametric priors over graphs, variational or MCMC Bayesian inference over weights, stochastic process formalisms for structure and scope, or the marriage of geometric deep learning (e.g., sheaf representations) with amortized variational inference. Bayesian GNNs achieve uncertainty quantification, model calibration, and improved robustness in cases of noisy, sparse, or partially observed graph data.

1. Foundational Bayesian Formulations in GNNs

The Bayesian formalism in GNNs arises by (i) treating the graph adjacency matrix, model parameters, or the message-passing process as latent random variables, and (ii) performing posterior inference conditioned on data. Distinct axes of "Bayesianization" have been studied:

  • Probabilistic Priors on Graphs: The adjacency AA is assumed unobserved or uncertain, with priors such as sparsity- and connectivity-promoting log-degree and Frobenius-norm penalties,

p(A)exp(α1log(A1)βAF2),A0,A=A,p(A) \propto \exp\left(\alpha\,\mathbf{1}^\top\log(A\mathbf{1}) - \beta\|A\|_F^2\right), \quad A \geq 0,\,A = A^\top,

and task-aware nonparametric likelihoods integrating node features and (if available) labels (Pal et al., 2020, Pal et al., 2019, Munikoti et al., 2020).

  • Bayesian Inference over Network Parameters: Standard GNN layers are endowed with Gaussian or dropout-based variational posteriors over their weights, optimized via ELBO:

LELBO=Eq(W)[logp(YX,G,W)]KL(q(W)p(W))\mathcal{L}_\text{ELBO} = \mathbb{E}_{q(W)}[\log p(Y|X, G, W)] - \text{KL}(q(W)\|p(W))

(Lamb et al., 2020, Mylonas et al., 2020, Komanduri et al., 2021).

  • Hierarchical/Beta-Process Priors: For flexible aggregation scope, e.g., in neighborhood adaptation, stick-breaking Beta-process priors are placed over layerwise inclusion probabilities, inferring the plausible number of hops for message passing (Regmi et al., 5 Feb 2026).
  • Structural Priors in Geometric Deep Learning: Bayesian models are imposed over structural objects such as sheaf restriction maps, using variational posteriors defined on matrix Lie groups (e.g., SO(d)SO(d) with reparameterizable Cayley distributions) (Gillespie et al., 2024).
  • Stochastic Generative Models over Interaction Graphs: In CF/recommender settings, ensemble-based generative modeling stochastically samples user-item graphs to propagate uncertainty through standard GNN layers (Gu et al., 2023).

2. Bayesian Inference Algorithms for BGNNs

BGNNs employ variational inference, MCMC, or stochastic process approximations depending on problem structure and computational constraints:

  • MAP Estimation in Nonparametric Models: Graph posterior modes are found via convex optimization, often with block-sparse or sparsity-inducing penalties, then fixed for downstream Bayesian GNN parameter inference with MC-dropout or Bayes-by-Backprop (Pal et al., 2020, Pal et al., 2019).
  • Monte Carlo and Variational Weight Posteriors: Local reparameterization, SGLD, SWAG, MC-dropout or factorized Gaussian approximate posteriors are used for network parameters. Fully factorized posteriors (mean-field) and more structured covariance models are both employed, typically maximizing ELBO amounts encompassing both regularization and task likelihood (Lamb et al., 2020, Mylonas et al., 2020, Komanduri et al., 2021).
  • Posterior over Structure and Parameters: Some models perform joint (but often decoupled or sequential) MAP estimation over the graph and Bayes-by-Backprop or MC-dropout over the weights (Pal et al., 2019, Munikoti et al., 2020). Other approaches jointly learn dynamic graph structure and GNN weights within a VAE/score-based framework (Sun et al., 2023).
  • Stochastic Process and Beta-Process Posteriors: In BNA (Regmi et al., 5 Feb 2026), the neighborhood scope per-layer is governed by a stick-breaking beta process, with variational parameters optimized via stochastic gradient and Concrete-Bernoulli reparameterization.
  • MCMC Posterior Inference over Hyperparameters: Explicit interpretable BNNs for structure learning are fit via HMC or NUTS over key thresholding/sparsity parameters, exploiting unrolled optimization steps that allow rapid joint sampling of all edge probabilities (Wasserman et al., 2024).

3. Modeling Uncertainty: Graph, Parameter, Structural, and Predictive

Bayesian GNNs produce uncertainty estimates at various modeling levels:

  • Graph Structure: Posterior predictive distributions over adjacency entries provide confidence in edge existence, yielding calibrated uncertainty bands, improved robustness under covariate shift (e.g., fewer signal samples or mismatched data-generating processes), and correlation between high uncertainty and misclassification (Wasserman et al., 2024).
  • Model Parameters: Posterior sampling over weights reduces overconfidence, especially out-of-distribution, and improves predictive calibration (miscalibration area, ECE). SGLD and Bayes-by-Backprop attain particularly low miscalibration area on molecular property prediction (Lamb et al., 2020).
  • Message-Passing or Structural Hyperparameters: In BNA (Regmi et al., 5 Feb 2026), uncertainty in the effective neighborhood hop is learned, leading to both improved prediction accuracy and well-calibrated predictions compared to fixed- or ensemble-hop schemes.
  • Latent Geometric Structure: In Bayesian sheaf neural networks, multiple sheaf samples per layer at inference yield reduced sensitivity to hyperparameters and improved accuracy on heterophilic graphs in the limited-data regime (Gillespie et al., 2024).

4. Representative Architectures and Datasets

Diverse BGNN architectures target a wide range of graph machine learning problems:

Model/Paper Probabilistic Latents Inference Main Task / Setting
Nonparametric BGCN (Pal et al., 2020, Pal et al., 2019) Adjacency matrix MAP + MC-dropout Node classification, link prediction, recommendation
BILGR (Munikoti et al., 2020) Graph + weights (MC-dropout on weights) MAP + MC Critical node identification
Bayesian MPNN (Lamb et al., 2020) Weights, readout GP, depth SGLD, BBP, SWAG Molecular property regression
BGCN-NRWS (Komanduri et al., 2021) Graph (via random walk), weights MCMC + VI Node classification
GNP (Carr et al., 2019) Predictive conditional distribution Maximum likelihood Edge imputation, graph regression
BSNN (Gillespie et al., 2024) Sheaf restriction maps (group-valued) Variational Node classification (heterophilic graphs)
BNA (Regmi et al., 5 Feb 2026) Per-layer beta-process masks Stochastic VI Node classification (homophilic, heterophilic), link prediction
GDBN (Sun et al., 2023) Sparse DAG adjacency, latent noise VAE Causal discovery in time series

These models span standard citation graphs (Cora, Citeseer, Pubmed), chemical molecules (QM9), synthetic and real dynamic networks, large-scale social/recommender systems, and physics-inspired sensor-graph simulations.

5. Evaluation Metrics, Calibration, and Empirical Results

Key evaluation metrics and findings in BGNN literature include:

  • Node/edge prediction accuracy, F1, AUC, SHD, and TPR/FDR: BGNNs generally outperform deterministic GNNs, especially in low-label and noisy regimes. For instance, BGCN-NRWS achieves substantial accuracy gains on Cora and Pubmed for 5-label/class semi-supervised splits (Komanduri et al., 2021), while non-parametric BGCN gains are pronounced for low-degree nodes (Pal et al., 2020).
  • Calibration metrics: Miscalibration area (MA) and expected calibration error (ECE) quantify predictive uncertainty versus true frequency. SGLD and Bayes-by-Backprop exhibit lower MA than MC-dropout and MAP baselines for molecular property prediction (Lamb et al., 2020). BNA drastically reduces ECE versus base GCNs (Regmi et al., 5 Feb 2026).
  • Robustness under noise and covariate shift: Integrated graph and parameter uncertainty yield empirical robustness to additional random edges and out-of-distribution test samples (Munikoti et al., 2020, Wasserman et al., 2024).
  • Posterior empirical coverage: In strain-based localization (Mylonas et al., 2020), MC posterior intervals achieve near-nominal empirical coverage.

6. Advanced Bayesian Geometric and Structured Models

BGNNs are extended to highly structured domains:

  • Sheaf-based GNNs: Bayesian sheaf neural networks variationally learn random restriction maps on SO(d)SO(d), with closed-form Cayley reparameterization and ELBO objectives achieving improved hyperparameter robustness and accuracy on heterophilic benchmarks (Gillespie et al., 2024).
  • Dynamic Bayesian networks via GNNs: GDBN uses a VAE with GNN-decoder and score-based L1 penalty to learn temporal causal adjacency, achieving superior F1 and AUC for non-linear VAR and real data compared to traditional methods (Sun et al., 2023).

7. Limitations, Open Problems, and Future Directions

Current limitations in Bayesian GNNs include:

  • Posterior Uncertainty Approximation: Many frameworks employ MAP point estimates or factorized posteriors for graphs and weights, thus underestimating uncertainty (e.g., lack of full variational p(A)p(A) in non-parametric BGCN (Pal et al., 2020, Pal et al., 2019)).
  • Computational Cost and Scalability: While many inference algorithms scale as O(nlogn)O(n\log n) or O(E)O(E) (with nn nodes, EE edges), the cost of MC sampling, marginalization over graphs, or posteriors over deep weights remains significant for very large graphs.
  • Expressivity and Over-smoothing: Uncertainty in aggregation scope (via learned priors, e.g., Beta-process), stochastic random walks, or adaptive connection sampling is necessary to mitigate over-smoothing but may complicate architectural design and stability (Regmi et al., 5 Feb 2026).
  • Lack of Fully Bayesian Treatments in Some Domains: For instance, ensemble-based graph sampling for recommendation (Gu et al., 2023) lacks closed-form ELBOs or Bayesian network priors on the interaction graph itself.

Proposed directions include scalable full Bayesian inference for graphs (e.g., via MCMC over AA), efficient variational (amortized) inference in dynamic and multi-relational graphs, extensions to hierarchical and temporal models, and deployment in uncertainty-sensitive applications in the sciences, finance, and engineering.


Key references: (Carr et al., 2019, Pal et al., 2019, Pal et al., 2020, Lamb et al., 2020, Mylonas et al., 2020, Munikoti et al., 2020, Komanduri et al., 2021, Wei et al., 2022, Sun et al., 2023, Gu et al., 2023, Wasserman et al., 2024, Gillespie et al., 2024, Regmi et al., 5 Feb 2026).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Bayesian Graph Neural Networks.