Papers
Topics
Authors
Recent
Search
2000 character limit reached

Radial Basis Operator Networks

Updated 2 April 2026
  • Radial Basis Operator Networks (RBONs) are models that extend classical RBF architectures to learn nonlinear operators between infinite-dimensional function spaces.
  • RBONs employ dual RBF layers—one applied to discretized inputs (branch) and one to query points (trunk)—to effectively approximate mappings in PDE and graph-based tasks.
  • Training methods for RBONs, including K-means-based initialization, least squares optimization, and gradient descent, ensure robust and efficient operator regression performance.

Radial Basis Operator Networks (RBONs) generalize classical radial basis function (RBF) architectures to operator learning, enabling the approximation of nonlinear operators that map between infinite-dimensional function spaces. These models leverage the expressivity of RBF representations and are structurally adapted for tasks such as scientific computing, PDE solution mapping, and operator regression on both vectorial, graph-based, and functional domains. The RBON paradigm encompasses pure operator approximators ("Radial-Basis Operator Networks"), adaptively parametrized operator nets rooted in Kolmogorov–Arnold theory, and graph-aware RBF networks for relational data.

1. Mathematical Formulation and Core Architecture

Let G:UVG:\mathcal{U}\to\mathcal{V} denote a continuous nonlinear operator acting between Banach spaces, frequently with UC(K1)\mathcal{U}\subset C(K_1) and VC(K2)\mathcal{V}\subset C(K_2). The modern RBON architecture approximates GG as GG^\dagger using a single hidden layer formed by the tensor product of two sets of RBFs: one ("branch") applied to discretized inputs um=(u(x1),,u(xm))Rmu^m=(u(x_1),\ldots,u(x_m))\in\mathbb{R}^m, and one ("trunk") applied to the query point yK2y\in K_2. The model's canonical form is

G(um)(y)=i=1Mk=1Nξik  φ(um;cib,σib)  φ(y;ckt,σkt)G^\dagger(u^m)(y) = \sum_{i=1}^M\sum_{k=1}^N \xi_{ik}\;\varphi(u^m;c_i^b,\sigma_i^b)\;\varphi(y;c_k^t,\sigma_k^t)

where φ(z;c,σ)=exp(zc2/(2σ2))\varphi(z;c,\sigma)=\exp(-\|z-c\|^2/(2\sigma^2)) is the Gaussian kernel, (cib,σib)(c_i^b,\sigma_i^b) and UC(K1)\mathcal{U}\subset C(K_1)0 are branch/trunk RBF centers and widths, and UC(K1)\mathcal{U}\subset C(K_1)1 are the linear output weights (Kurz et al., 2024).

In normalized RBONs (NRBON), the hidden layer activation is normalized by the UC(K1)\mathcal{U}\subset C(K_1)2 sum before output mapping. For operator learning on complex domains (e.g., frequency space), the same structure is applied with complex-valued inputs/centers and the complex-Euclidean norm.

2. Graph-based RBONs and Relational Data

RBONs can be adapted to graph-based settings where data is available only as a weighted adjacency matrix UC(K1)\mathcal{U}\subset C(K_1)3 encoding relations or metric-derived distances among UC(K1)\mathcal{U}\subset C(K_1)4 objects. The hidden layer constructs graph-RBF units as follows:

  • Define prototype weights UC(K1)\mathcal{U}\subset C(K_1)5, normalized to UC(K1)\mathcal{U}\subset C(K_1)6.
  • For node UC(K1)\mathcal{U}\subset C(K_1)7 and prototype UC(K1)\mathcal{U}\subset C(K_1)8, generalized "distance" is computed as

UC(K1)\mathcal{U}\subset C(K_1)9

  • The unit activation is VC(K2)\mathcal{V}\subset C(K_2)0.

This construction ensures that the RBON output is equivalent to a standard vector-RBF network whenever the adjacency matrix VC(K2)\mathcal{V}\subset C(K_2)1 encodes squared Euclidean distances between (potentially unobserved) latent vectors VC(K2)\mathcal{V}\subset C(K_2)2 (Sledge et al., 2019).

3. Training Methods and Parameter Selection

The training of RBONs proceeds via closed-form or gradient-based methods, depending on context:

  • For operator RBONs (Kurz et al., 2024):
    • The loss is the relative VC(K2)\mathcal{V}\subset C(K_2)3 error: VC(K2)\mathcal{V}\subset C(K_2)4.
    • The output weights VC(K2)\mathcal{V}\subset C(K_2)5 are solved for each trunk location via ordinary least squares or Moore–Penrose pseudoinverse. Centroids and widths are set using VC(K2)\mathcal{V}\subset C(K_2)6‑means clustering: branch RBF centers VC(K2)\mathcal{V}\subset C(K_2)7 are clusters of input samples; trunk centers VC(K2)\mathcal{V}\subset C(K_2)8 are clusters of query-points; widths are average intra-cluster radii.
    • Early stopping is applied to mitigate overfitting, especially for normalized variants.
  • For graph-RBONs (Sledge et al., 2019):

    • Parameters VC(K2)\mathcal{V}\subset C(K_2)9 are updated via gradient descent:

    GG0

    GG1 - Prototype parameters and RBF widths are updated by gradients directly on GG2. - All updates rely explicitly on entries of the adjacency matrix.

  • In adaptive RBF-based Kolmogorov–Arnold networks (Free-RBF-KAN), centroids and widths are fully trainable per layer; parameters are updated by backpropagation with constraints GG3 and GG4 restricted to the data domain for numerical stability (Chiu et al., 12 Jan 2026).

4. Universality and Approximation Guarantees

RBONs inherit the universal approximation property of RBF networks and extend it to operators:

  • If GG5 is a non-polynomial, rapidly-decaying RBF (e.g., Gaussian), then, for every continuous nonlinear operator GG6 and any GG7, there exist GG8 and network parameters such that

GG9

for the RBON mapping form above (Kurz et al., 2024). The proof extends Chen & Chen’s theorem to operator settings.

  • For Free-RBF-KANs that utilize the Kolmogorov–Arnold superposition theorem, expressive power is established by showing that networks composed of sums of univariate RBF subnets can uniformly approximate any continuous mapping GG^\dagger0 within arbitrary precision (Chiu et al., 12 Jan 2026).

A plausible implication is that RBONs offer rigorous universality in both function and operator spaces, although explicit error rates in Sobolev or Barron-type norms for operator settings remain an open problem.

5. Empirical Benchmarks and Comparative Analysis

Empirical results across canonical PDE operator learning tasks and standard regression benchmarks demonstrate that RBONs, including variants such as NRBON and frequency-domain RBON, offer strong generalization and compact representational efficiency, often outperforming paradigms such as DeepONet, LNO, and FNO (Kurz et al., 2024):

Network Wave (ID/OOD) Burgers (ID/OOD) Beam (ID/OOD)
RBON GG^\dagger1 GG^\dagger2 GG^\dagger3
NRBON GG^\dagger4 GG^\dagger5 GG^\dagger6

RBONs with a single hidden layer (≤225 units) exhibit low in- and out-of-distribution (OOD) errors, with OOD generalization robust even across function classes. In scientific data tasks (e.g., learning CO₂→temperature operators), RBONs match or improve on LSTM, FNO, DeepONet, and LNO baselines, especially in long-term forecasts (Kurz et al., 2024).

In the context of structured function learning, Free-RBF-KAN matches or surpasses classical B-spline KAN and standard RBF-KAN, providing reduced training/inference times and adaptive expressivity due to freely learnable centroids and widths (Chiu et al., 12 Jan 2026).

6. Architectural Variants: Kolmogorov–Arnold Free-RBF-KANs

Free-RBF-KAN combines Kolmogorov–Arnold superposition with univariate RBF expansions and adaptively learned centroids and widths, yielding a multilayer operator network architecture:

  • Layerwise, activations are

GG^\dagger7

with all parameters updated using joint backpropagation.

  • The architecture is embedded within operator learning pipelines, for example as trunk networks in DeepONet setups, consistently yielding lower relative GG^\dagger8 errors for comparable or reduced parameter counts.
  • Universality is guaranteed for continuous, non-polynomial RBF kernels (Chiu et al., 12 Jan 2026).

7. Limitations and Future Directions

  • RBON performance depends on the initialization of RBF centers and widths; GG^\dagger9-means clustering can be sensitive to local minima.
  • The size of the hidden layer grows quadratically with the number of branch/trunk RBFs (um=(u(x1),,u(xm))Rmu^m=(u(x_1),\ldots,u(x_m))\in\mathbb{R}^m0), impacting scalability for high-dimensional domains.
  • Current theoretical results provide only qualitative universality, with quantitative error rates a subject of ongoing research.
  • Exploration of alternative kernels beyond Gaussians (e.g., inverse-multiquadric) and sparse/low-rank tensor decompositions are open directions (Kurz et al., 2024).
  • Enhanced automation of RBF parameter initialization (e.g., density-based or ensemble clustering) and extension to multimodal and convolutional tasks for scientific and structured data are proposed avenues for future work.

RBONs establish a rigorous and flexible framework for operator learning, bridging RBF-based regression, functional approximation, and modern neural operator methods while maintaining computational tractability and compactness (Kurz et al., 2024, Sledge et al., 2019, Chiu et al., 12 Jan 2026).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Radial Basis Operator Networks (RBON).