Papers
Topics
Authors
Recent
Search
2000 character limit reached

JacobiConv: Spectral GNN with Jacobi Polynomials

Updated 4 March 2026
  • JacobiConv is a spectral GNN architecture that uses orthogonal Jacobi polynomial bases to achieve universal expressive power and fast convergence.
  • It adapts the filter representation to the empirical graph Laplacian spectrum via parameters α and β, improving Hessian conditioning and optimization stability.
  • The framework also enables fast Chebyshev–Jacobi transforms through stabilized recurrences and asymptotic formulas, ensuring numerical efficiency across diverse datasets.

JacobiConv is a spectral graph neural network (GNN) architecture that leverages Jacobi polynomial bases to parameterize graph filters. Originally proposed in the context of analyzing the expressive power and optimization landscape of spectral GNNs, JacobiConv abandons pointwise nonlinearities and instead focuses on orthogonal polynomial parameterizations, yielding models with provably universal expressive power under mild conditions and superior empirical performance on both synthetic and real-world datasets (Wang et al., 2022). The Jacobi polynomial basis is chosen for its ability to be adapted to the empirical distribution of graph Laplacian eigenvalues via parameters α and β, enabling accelerated convergence through improved Hessian conditioning in training. Separately, JacobiConv also refers to fast transforms between Chebyshev and Jacobi polynomials, realized through Hahn’s asymptotic formula and stabilized recurrences for numeric stability across the spectrum (Slevinsky, 2016).

1. Spectral GNNs and the Motivation for JacobiConv

In a spectral GNN framework, each node feature matrix XRn×dX\in\mathbb R^{n\times d} is interpreted as a graph signal to be filtered in the eigenspace of the graph’s normalized Laplacian L~\widetilde L. The action of a filter g()g(\cdot) on XX is formulated as Xg(L~)X=Ug(Λ)UXX \mapsto g(\widetilde L) X = U g(\Lambda) U^{\top} X, where UU is the eigenvector matrix and Λ\Lambda the diagonal matrix of eigenvalues (L~=UΛU\widetilde L = U\Lambda U^{\top}). To circumvent runtime eigendecomposition, gg is typically chosen polynomial in L~\widetilde L—classically g(L~)=k=0KθkL~kg(\widetilde L) = \sum_{k=0}^K \theta_k \widetilde L^k—and polynomial bases such as Chebyshev or Bernstein have been standard.

The choice of polynomial basis affects optimization: if gg is represented in a basis orthogonal with respect to the empirical spectral density of the graph signal, the Hessian in squared loss for filter coefficients is (nearly) diagonal, leading to better-conditioned gradients and faster convergence. Jacobi polynomials, parameterized by (α,β)>1(\alpha, \beta)>-1, provide a flexible family of orthogonal polynomials adjustable to the distribution of Laplacian eigenvalues encountered in real graphs (Wang et al., 2022).

2. Jacobi Polynomial Basis: Definitions and Properties

For k0k\ge0, Jacobi polynomials Pk(α,β)P_k^{(\alpha,\beta)} are defined on x[1,1]x\in[-1,1] via the three-term recurrence

P0(α,β)(x)=1,P1(α,β)(x)=12[(2+α+β)x+(αβ)],P_0^{(\alpha,\beta)}(x) = 1, \qquad P_1^{(\alpha,\beta)}(x) = \frac12[(2+\alpha+\beta)x + (\alpha-\beta)],

Pk(α,β)(x)=(Akx+Bk)Pk1(α,β)(x)CkPk2(α,β)(x),k2P_k^{(\alpha,\beta)}(x) = (A_k x + B_k) P_{k-1}^{(\alpha,\beta)}(x) - C_k P_{k-2}^{(\alpha,\beta)}(x),\quad k\ge2

with explicit Ak,Bk,CkA_k,B_k,C_k depending on α,β\alpha,\beta (see He et al. 2021; also (Slevinsky, 2016)). Jacobi polynomials are orthogonal with respect to the weight (1x)α(1+x)β(1-x)^{\alpha}(1+x)^{\beta}: 11(1x)α(1+x)βPm(α,β)(x)Pn(α,β)(x)dx=0 if mn.\int_{-1}^1 (1-x)^{\alpha}(1+x)^{\beta} P_m^{(\alpha,\beta)}(x) P_n^{(\alpha,\beta)}(x) dx = 0 \text{ if } m\ne n\,. In spectral GNNs, the normalized Laplacian spectrum [0,2][0,2] is shifted to [1,1][-1,1] via x1λx\mapsto 1-\lambda, and the Jacobi polynomial basis Pk(α,β)(IL~)P_k^{(\alpha,\beta)}(I-\widetilde L) is used as the functional basis for gg.

3. JacobiConv Architecture and Spectral Filter Parameterization

JacobiConv parameterizes the spectral filter gg as a KKth-order Jacobi polynomial expansion: gθ(L~)=k=0KθkPk(α,β)(L~),g_\theta(\widetilde L) = \sum_{k=0}^K \theta_k P_k^{(\alpha,\beta)}(\widetilde L), where the θk\theta_k are learned coefficients, and each output channel \ell may use its own {θk()}\{\theta_k^{(\ell)}\}. In the forward computation, the input XX is linearly projected by WRd×dW\in\mathbb R^{d\times d'} to XWXW, followed by spectral filtering to produce Z=gθ(L~)XWZ = g_\theta(\widetilde L) XW.

Filter learning is cast as minimizing Frobenius loss R(θ,W)=gθ(L~)(XW)YF2R(\theta,W) = \|g_\theta(\widetilde L)(XW) - Y\|_F^2 with 2\ell_2 weight decay on both WW and the filter coefficients. Optimization is performed via Adam with grid/random search for learning rates and polynomial order K[5,20]K\in[5,20].

To further stabilize coefficient learning, JacobiConv employs Polynomial-Coefficient-Decomposition (PCD), expressing θk\theta_k as bounded linear combinations θk=i=1Byinik\theta_k = \sum_{i=1}^B y_i n_{ik} with yiy_i bounded by tanh\tanh nonlinearity.

4. Orthogonality, Hessian Conditioning, and Adaptivity

Near a loss-minimizing solution, the Hessian with respect to the filter coefficients θ\theta is determined by the inner products gk(x)g(x)f(x)dx\int g_k(x)g_\ell(x)f(x)dx, where f(x)f(x) is the empirical spectral density weighted by (UX)i2|(U^\top X)_i|^2 (Fourier coefficients of XX). Orthogonalizing the basis gkg_k for this ff minimizes the Hessian's condition number. Jacobi polynomials' orthogonality to (1x)α(1+x)β(1-x)^{\alpha}(1+x)^{\beta} provides the flexibility to fit the observed empirical density f(x)f(x) by grid-searching over α,β[1,2]\alpha,\beta\in[-1,2].

This adaptivity is not present in fixed bases such as Chebyshev or Bernstein, allowing JacobiConv to maintain fast convergence and stable optimization over diverse spectral densities encountered in practice (Wang et al., 2022).

5. Expressive Power and Universality without Nonlinearities

Theorem 4.1 of (Wang et al., 2022) shows that a purely linear spectral GNN of the form Z=g(L~)XWZ=g(\widetilde L) X W can realize any mapping XZX\mapsto Z provided:

  • L~\widetilde L has distinct eigenvalues (no repeated eigenvalues).
  • XX has no missing frequency components (nonzero in every eigen-direction).

Consequently, adding nonlinearities does not increase expressive power for general graphs with suitable features. Nonlinearities can assist in degenerate edge cases (repeated eigenvalues, missing frequencies), but these are empirically rare (<1%<1\% repeated eigenvalues). Universality extends to multi-output nodes by assigning each output channel an independent g()g^{(\ell)}. Further, polynomial filter GNNs have at most the discriminative power of (K+1)(K+1)-step 1-Weisfeiler–Leman (1-WL), which in the absence of eigenvalue multiplicity and frequency gaps already achieves full node distinction.

6. Empirical Results and Benchmark Comparisons

JacobiConv's empirical validation covers both synthetic filtering tasks and real-world graph classification:

Task/Domain Baselines JacobiConv outcome
Synthetic filtering ("image-on-graph") GPRGNN, ARMA, ChebyNet, BernNet, monomial/Chebyshev/Bernstein/Jacobi linear GNNs Up to 50× lower MSE (e.g., 3×1043\times10^{-4} for Jacobi vs. $1.8$ for ARMA on low-pass); outperforms all linear baselines (10× lower than monomial, Bernstein)
Real-world node classification (10 datasets) GCN, APPNP, ChebyNet, GPRGNN, BernNet (with nonlinearities) Wins 9/10 datasets; up to +12% accuracy gain (e.g., Squirrel); 2–3 points better than BernNet on average, using 10%\sim 10\% as many parameters

This demonstrates that JacobiConv delivers universal spectral filtering capability and state-of-the-art empirical performance using only linear operations and no activations (Wang et al., 2022).

7. Fast Chebyshev–Jacobi Transforms and Numerical Implementation

In the context of polynomial basis transforms, JacobiConv also refers to the fast numerically stable computation of Chebyshev–Jacobi transforms (Slevinsky, 2016). This is realized using:

  • Hahn’s interior asymptotic formula for Pn(α,β)(x)P_n^{(\alpha,\beta)}(x) with rigorous error bounds, allowing reduction to a sum of diagonally scaled DCT-I and DST-I transforms for "asymptotic blocks".
  • Stable three-term recurrence and the Clenshaw–Smith algorithm, with Reinsch's endpoint modifications to maintain uniform accuracy even near x=±1x=\pm1.
  • Complexity O(Nlog2N/loglogN)O(N\log^2N/\log\log N) for the full transform, using M58M\approx5-8 for double precision and careful block partitioning of the computation domain.

Key implementation details include:

  • Pre-planning DCTs/DSTs via FFTW or equivalent libraries.
  • Explicitly handling edge cases and parameter regimes for α,β\alpha,\beta, including parameter shifting for half-integer cases.
  • Clenshaw–Curtis quadrature and accurate endpoint handling (Slevinsky, 2016).

This fast transform is essential for efficient evaluation and inversion of Jacobi polynomial expansions in GNN filtering and beyond.


In conclusion, JacobiConv unifies advances in spectral graph filtering, polynomial basis adaptivity, and fast, stable transforms to provide a highly expressive, optimizable, and empirically dominant spectral GNN framework without the need for nonlinearities or large overparameterization (Wang et al., 2022, Slevinsky, 2016).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to JacobiConv.