Physics-Informed Chebyshev Neural Operator
- CPNO is a mesh-free operator learning framework that leverages Chebyshev polynomial bases for enhanced stability and spectral accuracy in solving parameterized PDEs.
- It incorporates a parameter-dependent modulation mechanism within a deep neural network to effectively integrate heterogeneous, multi-scale, and high-frequency PDE data.
- Empirical results on benchmarks demonstrate state-of-the-art performance with rapid convergence and reduced errors in complex scenarios such as transonic airfoil flows.
The Physics-Informed Chebyshev Polynomial Neural Operator (CPNO) is a mesh-free deep learning framework devised to solve parameterized partial differential equations (PDEs). It replaces monomial or Fourier-type feature expansions with Chebyshev polynomial bases, thereby ensuring enhanced numerical stability, spectral convergence, and robustness in physics-informed learning settings. CPNO introduces a parameter-dependent modulation mechanism to seamlessly integrate heterogeneous PDE data, making it particularly effective for multi-scale, high-frequency, and parametric PDE operator learning, including applications in complex geometries such as transonic airfoil flow (Chen et al., 2 Feb 2026).
1. Chebyshev Polynomial Expansions and Functional Representation
CPNO employs the Chebyshev polynomials of the first kind, , for stable and efficient representation of functions defined on . Formally,
These polynomials are orthogonal under the weight , and satisfy a uniform bound , making them suitable for constructing spectral approximations. For a sufficiently smooth function ,
with coefficients given by a weighted inner product. For multi-dimensional domains (including time and parameters), tensor-product bases are used, ensuring that feature representations encompass each spatial/temporal and parametric input (Chen et al., 2 Feb 2026, Mostajeran et al., 6 Jan 2025).
2. CPNO Architectural Design
CPNO’s core is a -layer synthesis network that recursively builds polynomial-valued feature maps. The initial feature vector is a concatenation of Chebyshev features evaluated coordinate-wise (inputs normalized to ). Each subsequent layer performs
where is a learnable linear spectral projection, is Hadamard (elementwise) product (effectively raising polynomial degree), and are parametric modulation vectors output by mapping networks ingesting PDE parameters . The nonlinearity is set to GELU or tanh to ensure smoothness and spectral fidelity. After layers, a final affine map outputs the PDE solution or its expansion coefficients.
Pseudocode summarizing the network forward pass:
1 2 3 4 5 6 7 8 9 10 |
x_, t_ = normalize(x, t) # to [-1, 1] Φ = ChebyshevFeatures(x_, t_) h = Φ for j in range(q): p = A_j @ Φ s = p * h ω, φ = MappingNet_j(θ) h = GELU(ω * s + φ) u_pred = W_q @ h + b_q return u_pred |
3. Physics-Informed Loss Formulation
The network is trained by minimizing a composite loss:
- PDE Residual:
- Initial Condition:
- Boundary Condition:
- Data Fit (optional):
Hyperparameters balance the loss components; typical values set PDE, initial, and boundary terms to $1.0$, data term (when used) to (Chen et al., 2 Feb 2026).
4. Theoretical Properties and Numerical Conditioning
CPNO’s use of the Chebyshev basis addresses two critical operator learning challenges:
- Spectral Bias Mitigation: The Chebyshev expansion exhibits near-minimax uniform approximation error. For analytic on a Bernstein ellipse, the best -term Chebyshev truncation satisfies
for some and constant .
- Stability and Conditioning: The Lebesgue constant for Chebyshev nodes grows only logarithmically,
and the condition number of the Chebyshev-gram matrix , dramatically lower than for monomials. This ensures stable gradient propagation and robust training even at high polynomial orders (Chen et al., 2 Feb 2026, Mostajeran et al., 6 Jan 2025).
5. Benchmark Performance and Empirical Results
CPNO achieves state-of-the-art accuracy and convergence rates on a wide range of parameterized PDE benchmarks:
- Burgers’ equation: error (zero-shot, no data), outperforming PI-DeepONet () and HyperPINNs ().
- Allen–Cahn equation: (zero-shot).
- Diffusion–Reaction: (zero-shot).
- 2D vorticity–Navier–Stokes: (zero-shot).
- Few-shot regime: Additional three solution snapshots reduce errors further (e.g., for Allen–Cahn).
The framework demonstrates rapid convergence: errors are typically attained in less than 5000 epochs, whereas baseline neural operator models require over 20,000 epochs. In frequency analysis, CPNO captures high-wavenumber content () not resolved by MLP-based alternatives.
In the transonic airfoil flow experiment (parameterized complex geometry), CPNO with Chebyshev order 10, network depth 12, and degree-16 encoding achieves L1 errors on of mean and max , demonstrating its capacity for operator learning on challenging fluid dynamics problems (Chen et al., 2 Feb 2026).
6. Implementation Best Practices
Recommended configuration for CPNO includes:
- Chebyshev polynomial order for smooth fields; for sharp layers or shocks.
- Depth layers; hidden dimension (typically $64–128$).
- Activation: GELU or tanh; ReLU is discouraged due to nondifferentiability affecting spectral accuracy.
- Adam optimizer with learning rate warmed up to and multiplicative decay.
- Minibatch of $2000–4000$ uniformly-sampled collocation points per step.
- Normalization of each physical domain to to ensure polynomial features remain well-scaled.
- Mapping networks for parameter-dependent modulation are kept shallow and narrow (2–3 layers, width 16–32) to control overfitting (Chen et al., 2 Feb 2026).
7. Extensions, Limitations, and Context
CPNO’s foundation explicitly incorporates the Chebyshev basis, parameter modulation, and mesh-free operator learning, distinguishing it from monomial-MLPs, Fourier-based, or vanilla Galerkin and KAN variants (Guo et al., 2024, Zhang et al., 13 May 2025, Mostajeran et al., 6 Jan 2025). While CPNO is robust to multi-scale and high-frequency phenomena, practical limitations include the expense of high-dimensional tensor-product expansions and domain normalization strategies for non-rectilinear geometries. Potential future work involves sparse-grid Chebyshev representations, adaptive network depth, and hybridization with attention-based enhancements as in recent Chebyshev–KAN models (Zhang et al., 13 May 2025).
CPNO situates itself within the rapidly evolving landscape of neural operator design, providing a theoretically principled and empirically validated approach to solving parametric, time-dependent, and nonlinear PDEs with superior accuracy, efficiency, and training robustness (Chen et al., 2 Feb 2026).