Bayesian Interpolating Neural Network
- B-INN is a probabilistic surrogate modeling framework that extends 1D interpolating networks to multi-dimensional domains using tensor algebra and Bayesian inference.
- The model employs a block-alternating least squares algorithm with closed-form Bayesian updates to achieve efficient, linear-scaling uncertainty estimation.
- Empirical results demonstrate that B-INN delivers competitive accuracy and substantial speedups over traditional Gaussian processes and Bayesian neural networks in active learning settings.
The Bayesian Interpolating Neural Network (B-INN) is a probabilistic surrogate modeling framework designed to address scalability and reliability challenges in uncertainty quantification for large-scale, industry-driven simulations. The B-INN combines high-order interpolation, tensor decompositions, and alternating direction methods to achieve data efficiency and robust uncertainty estimation, enabling practical active learning in high-dimensional, data-intensive physical systems (Park et al., 30 Jan 2026).
1. Mathematical Formulation and Model Architecture
The Bayesian Interpolating Neural Network is founded on the extension of 1D Interpolating Neural Networks (INN) to multidimensional domains through tensor algebra. In the one-dimensional case, a scalar function is expressed as a linear combination of fixed interpolation basis functions with trainable weights: where each is selected for numerical qualities such as compact support and smoothness.
For inputs in dimensions, a na\"ive full grid combination is computationally infeasible for large . Instead, the B-INN uses a rank- CANDECOMP/PARAFAC (CP) tensor decomposition: with . This architecture can be interpreted as a shallow neural network with a single hidden layer of neurons, each neuron computing the product across dimensions.
The block-alternating scheme freezes weights in dimensions to focus the inference or update step on a single dimension, yielding a design matrix and allowing efficient iterative optimization.
2. Bayesian Inference and Alternating Least Squares
Transitioning from the interpolating neural network to its Bayesian instantiation, independent spherical Gaussian priors are imposed on all weights per block:
Given the data, Bayesian linear regression is performed dimension-wise: The posterior for is available in closed form: where
The block-wise alternating update cycles through dimensions, each time reconstructing with updated means in other subspaces. Convergence is achieved by iterating this procedure for a fixed number of Alternating Least Squares (ALS) steps.
3. Connection to Gaussian Processes
The B-INN’s function space is contained within that of Gaussian processes (GPs). In the limit as the mode rank , and with normalization , the prior over functions converges (via the multivariate Central Limit Theorem) to a Gaussian process with kernel: With Gaussian additive noise, the finite-dimensional posteriors of B-INN converge to those of GPs with kernel as . Thus, B-INN provides a tractable, low-rank surrogate whose function class approaches that of GPs in the infinite-rank regime.
4. Complexity, Predictive Formulas, and Algorithm
After block-wise regression, predictions at a test input use the closed-form mean: The variance (separating epistemic and aleatoric contributions) is mode-wise: with denoting the vector of interpolation basis evaluations for dimension .
The computational complexity per ALS sweep is , with iterations and dimensions, crucially linear in sample size given that . For comparison, standard GPs require operations.
Training Algorithm:
- Initialize all to zero.
- For each ALS iteration and each dimension :
- Compute frozen factors for all .
- Build design matrix .
- Solve Bayesian linear regression (BLR): obtain .
- Update weights .
- Return for all . Inference of predictive mean and variance follows the closed formulas above.
5. Empirical Performance and Benchmark Results
B-INN demonstrates superior computational and statistical performance in surrogate modeling tasks:
- 1D Regression (Synthetic): For with noise, B-INN matches GP accuracy () using basis at lower computation, while BNNs present higher error and significant sampling overhead. Training time for B-INN with is approximately 1s (– faster than fastest BNNs).
- Aerodynamic Surrogate (BlendedNet 7→4):
| Model | RMSE | Training Time (s) | |---------------|----------|---------------------| | B-INN | 0.02–0.05| 20–30 | | BNN-HMC | 0.02–0.05| 575 | | BNN-VI | 0.05–0.08| 550–610 |
B-INN achieves comparable or better accuracy at least faster.
- Active Learning on PDEs:
- Poisson (3D+1 parametric): B-INN initial RMSE , best RMSE after AL (training time 261s), compared to BNN-HMC (/, 8233s) and BNN-VI (/, 19321s).
- Heat Eq (2D+time+2 param): B-INN initial RMSE , best , training time 7297s; BNN-HMC /, 32984s; BNN-VI /, 91960s.
B-INN yields 10–50 speedups with lower error and more robust uncertainty than BNN alternatives.
6. Large-Scale Active Learning and Applications
B-INN is engineered for workflows where training data is acquired at great cost (e.g., industrial simulations with high-fidelity physics engines). Its retraining complexity, closed-form uncertainty quantification, and tensorized basis enable practical active learning loops at scales unattainable by full GPs or standard BNNs.
B-INN’s calibrated epistemic variance quantifies model uncertainty, directly informing acquisition of new data points. This enables error reductions of approximately 90% within tens of active learning rounds, as opposed to hundreds required by less data-efficient architectures. Variational BNNs may systematically underestimate uncertainty, resulting in poor acquisition and slow error reduction. By contrast, B-INN maintains uncertainty calibration across scales, supporting robust design and simulation.
7. Theoretical and Practical Significance
The B-INN embodies a hybrid point between classical interpolation, low-rank tensor methods, and modern Bayesian inference. Its function class, as , is a proper subset of GPs, allowing rigorous theoretical comparison. Linear scaling with sample size and tractable posterior updates distinguish it from other surrogate paradigms in both theory and practice.
B-INN constitutes a practical, well-calibrated foundation for large-scale, uncertainty-driven modeling, simulation, and design, particularly where rapid retraining and reliable uncertainty quantification are vital (Park et al., 30 Jan 2026). Its empirical speedups (20–10,000 over GPs/BNNs) and accuracy have direct implications in computational physics, engineering design, and scientific computing contexts where data/compute bottlenecks are prevalent.