Efficient Curvature-Aware Graph Network
- The paper introduces an Efficient Curvature-aware Graph Network that replaces computationally heavy Ollivier–Ricci curvature with Effective Resistance Curvature (ERC) to achieve fast, scalable graph processing.
- It leverages the Laplacian pseudoinverse for effective resistance computation, offering speedups of 60×–1300× compared to traditional optimal transport methods.
- ERC enhances curvature-aware message passing in GNNs, improving node and graph classification while maintaining interpretability and robustness across various graph conditions.
Efficient Curvature-aware Graph Network models exploit graph curvature as a geometric prior to enhance the representation power, robustness, and interpretability of Graph Neural Networks (GNNs). Traditional curvature-aware GNNs use Ollivier–Ricci curvature for its geometric expressiveness, but its reliance on optimal transport computations results in prohibitive preprocessing cost for large-scale graphs. The Efficient Curvature-aware Graph Network advances the state of the art by introducing Effective Resistance Curvature (ERC), a scalable measure based on the Laplacian pseudoinverse and effective resistances, which maintains near-equivalent capacity to Ricci curvature in modeling local geometric structure but drastically reduces computational overhead (Fei et al., 3 Nov 2025).
1. Mathematical Foundations: Effective Resistance Curvature
Let be an undirected graph with weighted adjacency matrix and Laplacian , where . The Moore–Penrose pseudoinverse defines the effective resistance between nodes and : where is the th basis vector. For each edge , the relative resistance is
The node-resistance curvature: and the edge-resistance curvature: Theoretical analysis demonstrates that under small random walk time scales, ERC and Ollivier–Ricci curvature coincide to first order and share monotonicity properties: increasing weights or adding edges always increases curvature on adjacent edges. Proper normalization gives and establishes substitutability for practical applications.
2. Computational Efficiency and Complexity
The bottleneck in Ricci curvature is solving an optimal transport (Wasserstein) problem per edge—a process with worst-case complexity on graphs of nodes. For ERC, the cost is dominated by Laplacian inversion, for dense graphs and for sparse graphs, and each resistance becomes a simple quadratic form evaluation. On massive graphs, ERC permits further acceleration by Cholesky, conjugate gradient, or low-rank Laplacian sketch techniques. Real benchmarks report CPU ERC computation times of $0.04$–$6$ seconds versus $3$–$900$ seconds for Ricci curvature (speedups –).
3. Algorithmic Implementation
A typical ERC computation pipeline is as follows:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 |
Input: graph Laplacian L, regularization parameter ε L̄ = L + ε*I M = inverse(L̄) # For small n, direct; for large n, use sparse linear solver for each edge (i,j) in E: d = e_i - e_j R_ij = d.T @ (M @ d) # Or use conjugate gradient per d ω_ij = c_ij * R_ij for each node i: p_i = 1 - 0.5 * sum over neighbors j of ω_ij for each edge (i,j): k_ij = 2*(p_i + p_j)/R_ij |
ERC values are then incorporated as edge weighting functions in the message-passing layers.
4. Curvature-aware Message Passing and Network Architecture
ERC is used to modulate each message via a curvature weighting function (examples: , or ):
In matrix form: where .
For node classification, is read out per node; for graph classification, a global pooling/sum is applied followed by an MLP.
5. Training Protocol and Hyperparameter Settings
Loss functions include standard cross-entropy for classification and MSE for regression. Optionally, a curvature-smoothness regularizer: The recommended protocol is:
- Hidden dimensions: 64
- Layers: 2–4 graph convolution layers
- Learning rate: 0.005 (Adam)
- Weight decay:
- Dropout: 0.5
For large graphs, mini-batch sampling is used. ERC values are precomputed and stored as sparse edge attributes.
6. Empirical Evaluation and Comparative Results
ERC-GNN matches or modestly exceeds OR-GNN accuracy (mean absolute difference ≈ 0.16%). Representative results:
- Node classification: Cora, Citeseer, PubMed, Amazon, Coauthor, synthetic benchmarks
- Graph classification (global pooling): ENZYMES, MUTAG, PROTEINS, D&D, IMDB-B, COLLAB
ERC computation time is between $0.04$ and $6$s (vs. $3$–$900$s for Ricci curvature), and memory footprint is minimal (no intermediate transport plans). Ablations show ERC and Ricci curvature boost performance over standard GNNs; in high-density regimes, Ricci curvature may outperform ERC under certain schemes, but tuning can recover or exceed the difference. Distributional analysis indicates ERC captures more extreme geometric anomalies via its heavier-tailed curvature distribution.
7. Practical Considerations, Robustness, and Limitations
ERC is recommended for static graphs of moderate or large size, with computation best performed offline using sparse linear algebra. It is robust to regularization () and random graph perturbations. With dynamic/evolving graphs, efficient incremental updates of ERC remain open. For directed or highly heterogeneous graphs, generalizations of ERC are required, as is end-to-end learning of functions. Non-small- interpretations await further theoretical development.
ERC offers a scalable, interpretable geometric prior for GNNs, enabling curvature-aware message passing on previously intractable graph sizes, while preserving the accuracy and robustness of Ricci curvature approaches (Fei et al., 3 Nov 2025).
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free