Compositional Neural Operators (CompNO)
- Compositional Neural Operators (CompNO) are a modular framework that decomposes the evolution of PDEs into reusable neural blocks corresponding to fundamental differential operators.
- The method pretrains separate Fourier Neural Operators for elementary tasks like convection, diffusion, and nonlinear convection, then assembles them for coupled PDE solutions.
- Empirical results show that CompNO achieves superior data efficiency, reduced parameter counts, and enhanced physical interpretability compared to monolithic architectures.
Compositional Neural Operators (CompNO) are a modular framework for scientific foundation modeling of parametric partial differential equations (PDEs), designed to overcome the computational and interpretability limitations of monolithic neural operator architectures. By decomposing the learning of evolution operators for complex PDE systems into reusable neural blocks that correspond to fundamental differential operators, CompNO enables efficient pretraining, flexible assembly, and rigorous enforcement of boundary conditions. This approach has demonstrable advantages in data efficiency, physical interpretability, and adaptability across diverse parameter regimes.
1. Framework Overview and Architectural Components
CompNO is formulated under a “pretrain–assemble–fine-tune” paradigm for foundation modeling of parametric PDEs (Hmida et al., 12 Jan 2026). The architecture consists of three key components: Foundation Blocks, Adaptation Blocks, and Boundary-Condition Operators.
Foundation Blocks
Each Foundation Block is a parametric Fourier Neural Operator (FNO), trained to approximate the temporal evolution for a specific elementary differential operator—such as convection, diffusion, or nonlinear convection. The FNO paradigm maps functions via truncated Fourier transforms, channel-wise linear mappings , nonlinearities , and inverse Fourier transforms: To handle PDE parameters (e.g., velocity , viscosity ), CompNO appends these parameters as extra channels, constructing parametric input features and propagating these through stacked Fourier layers, each with tensorized spectral mixing and skip connections.
Specialized Foundation Blocks
- Convection Block: Trained on with sampled from .
- Diffusion Block: Trained on , .
- Nonlinear Convection Block: Trained on .
Each block is pretrained independently and its parameters are frozen after pretraining.
Adaptation Blocks (Aggregator)
Adaptation Blocks are shallow task-specific neural networks that assemble one or more Foundation Blocks into a solver for a coupled PDE: For linear systems (e.g., convection-diffusion), may be a linear aggregator: with learned to minimize the PDE residual. For nonlinear systems (e.g., Burgers’ equation), a nonlinear MLP aggregator is necessary to represent cross-terms that cannot be captured by linear combinations.
Boundary-Condition Operator
A dedicated Boundary-Condition Operator enforces Dirichlet boundary values exactly at inference time, employing the BOON procedure. This involves boundary response computation of the FNO kernel, pre-correction of values at endpoints, and explicit overwriting to guarantee zero discrepancy with prescribed boundary data. Unlike baseline neural operators (e.g., PFNO, PDEFormer), this ensures exact satisfaction of boundary constraints in all experiments.
2. Training Regime, Losses, and Evaluation Metrics
The CompNO pipeline separates pretraining of Foundation Blocks from fine-tuning of Adaptation Blocks:
- Foundation Block Pretraining: Each block is trained for 1000 epochs (batch size 50) with Adam (initial learning rate , StepLR every 100 epochs) to minimize mean squared error (MSE) loss:
Training data includes 8,000 instances per operator; 2,000 are reserved for testing.
- Adaptation Block Fine-tuning: With Foundation Blocks' weights frozen, only aggregator parameters are optimized for 100 epochs (batch size 4, Adam, learning rate ). Mean absolute error (MAE) is used:
MAE is selected over MSE to stabilize gradients in few-shot adaptation regimes.
- Evaluation Metric: Relative error is used to assess prediction accuracy:
or, time-averaged with normalization.
3. Numerical Results and Empirical Comparison
CompNO was evaluated on the PDEBench one-dimensional suite, including convection, diffusion, convection–diffusion, and Burgers’ equations. The comparison metrics include test MAE, MSE, and relative errors; baseline models include PFNO, PDEFormer, and in-context learning (ICL) based models.
Pretraining Quality
| Foundation Block | Parameters | MSE train () | MSE test () |
|---|---|---|---|
| Convection PFNO | 217k | 0.95 | 0.98 |
| Diffusion PFNO | 217k | 5.7 | 5.8 |
| Nonlinear-Convection FNO | 75k | 7.6 | 8.0 |
Aggregator Adaptation (Coupled PDEs)
| Problem | CompNO Params | CompNO Train MAE | CompNO Test MAE () | Boundary Error | PFNO Params | PFNO Train MAE | PFNO Boundary Error |
|---|---|---|---|---|---|---|---|
| Convection–Diffusion | 257 | 0.008 | 0.019 | 0 | 217k | 0.039 | 1.53 |
| Burgers’ | 132k | 0.01 | 0.03 | 0 | 217k | 0.002 | 0.01 |
Relative Test Error (PDEBench)
| PDE | PDEFormer-FS | ICL-based | CompNO |
|---|---|---|---|
| Convection | 0.0124 | 0.0184 | 0.0071 |
| Diffusion | — | 0.0120 | 0.0075 |
| Convection–Diffusion | — | 0.0231 | 0.0120 |
| Burgers () | 0.0135 | — | 0.0280 |
| Burgers () | 0.0399 | — | 0.0371 |
4. Generalization, Boundary Satisfaction, and Interpretability
CompNO demonstrates robust generalization to broader parameter regimes. For convection–diffusion, it achieves the best accuracy in diffusion-dominated, low-Peclet-number flows (Pe 2). Error increases in advection-dominated regimes but remains competitive. For Burgers’ equation, test error is stable for low Reynolds numbers (Re 1), peaks in the transition regime (Re 10–100), and partially recovers at high Re, indicating an ability to resolve large-scale structures but only approximate rapid inertial features.
Boundary satisfaction is exact for Dirichlet conditions; boundary error is zero by construction in all experiments, which further reduces interior prediction error (MAE decreases from 0.025 to 0.004 when BCs are enforced). This distinguishes CompNO fundamentally from baselines, for which endpoint mismatches occur. The model trained on a $1024$-point grid generalizes to $2048$ points with only moderate MAE increase, confirming operator- rather than grid-based learning.
Interpretability is enhanced because each Foundation Block corresponds to a known mathematical operator (such as , , or a nonlinear term), and the aggregation weights directly reflect a physics-informed decomposition of the target evolution operator.
5. Data Efficiency, Scalability, and Limitations
CompNO is highly data-efficient: Foundation Blocks only require pretraining on elementary operators, which is computationally cheap. Adaptation Blocks are amenable to few-shot fine-tuning since they comprise only a handful of parameters for linear aggregation or small MLPs for nonlinear aggregation. This yields substantial parameter count reduction compared to monolithic networks: e.g., CompNO-v1 for convection–diffusion with 257 parameters versus PFNO's 217k.
The architecture is mesh-invariant and dimension-agnostic, due to its reliance on spectral FNOs, suggesting direct extensions to 2D and 3D PDEs. Nonetheless, such extensions may require more expressive aggregation schemes and new boundary-enforcement strategies.
All reported experiments are one-dimensional. Current adaptation blocks may underfit in scenarios with strong nonlinearity or developed turbulence, especially as Pe or Re increases. A plausible implication is that performance on turbulent or high-dimensional flows may be bottlenecked by the expressivity of both the operator library and the chosen aggregator.
6. Future Directions and Research Outlook
Planned improvements for CompNO include enriching the Foundation Block library to encompass additional operators (e.g., pressure-gradient, divergence), exploring hierarchical or variational aggregator architectures, and systematically validating the pipeline on multi-dimensional Navier–Stokes systems and complex geometries. Such developments are intended to establish CompNO as a general-purpose PDE foundation model, scalable across domains and adaptable to both linear and strongly nonlinear systems (Hmida et al., 12 Jan 2026).