Memory-Informed Neural Pseudo-Operator (MINPO)
- MINPO is a mesh-free neural framework that models nonlocal spatiotemporal dynamics and fractional PDEs via learned memory operators and inverse reconstructions.
- It integrates MLP and Kolmogorov-Arnold networks as encoders with a nonlocal-consistency loss to enforce coherence between the learned operator and the reconstructed dynamics.
- Benchmark experiments show MINPO achieves significant error reductions and speedups compared to classical solvers and existing PINN approaches.
The Memory-Informed Neural Pseudo-Operator (MINPO) is a mesh-free neural framework designed to model and resolve nonlocal spatiotemporal dynamics described by integro-differential equations (IDEs). It encapsulates both the memory operator and its inverse through neural networks, enabling explicit reconstruction of unknown solution fields. The architecture leverages either multilayer perceptrons (MLPs) or Kolmogorov-Arnold Networks (KANs) as encoders and introduces a nonlocal consistency loss to enforce coherence between the learned operator field and the reconstructed dynamics. MINPO is applicable to systems governed by diverse nonlocal kernel structures, including fractional partial differential equations (PDEs), and demonstrates robust generalization across problem classes and kernel dimensionalities (Mostajeran et al., 19 Dec 2025).
1. Mathematical Formulation of Nonlocal Dynamics
MINPO targets the general form of an IDE embodying spatial nonlocality, temporal memory, and fractional dynamics. The governing equation in space-time coordinates is:
- : local (possibly nonlinear) differential operator
- : composite time-evolution operator, including the first-order and fractional Caputo derivative,
where is the Caputo derivative,
- : unified nonlocal memory operator, for general kernel ,
where denotes the causal past of .
The inverse of the Caputo derivative is the Riemann–Liouville fractional integral:
Many classical and fractional nonlocal PDEs are subsumed as special cases of this general structure.
2. Neural Architecture and Operator Representation
MINPO learns the memory operator and its inverse via two neural networks:
- Memory encoder
- Inverse-memory encoder (for fractional dynamics)
Encoder Choices
- MLP encoder: three hidden layers, 33 neurons per layer with activation
- KAN encoder: three hidden layers, 15 neurons, each neuron is a tunable Chebyshev-polynomial map (degree ), with nonlinearities between layers
KANs simultaneously learn linear weights and coefficients of univariate Chebyshev activation maps, yielding enhanced parameter efficiency for nonlocal operator learning.
Explicit Solution Reconstruction
Given (and in the fractional case), MINPO reconstructs the solution via an explicit differentiable ansatz:
- indexes spatial/time derivatives
- Coefficients arise from applying Leibniz' rule (or corresponding identities for fractional inversion)
- enforces initial traces or fractional inverse requirements
Common reduction cases:
| Experiment | Solution Reconstruction |
|---|---|
| Volterra IDE (Exp I) | |
| 3D nested IDE (Exp II) | |
| Fractional PDE (Exp III) |
3. Loss Functions and Training Procedure
All neural parameters are jointly optimized using a composite loss function:
- Physics-residual loss (): measures MSE residual over collocation points,
Crucially, no kernel or fractional term is discretized in this residual; all nonlocality resides in and .
- Data-fidelity loss (): penalizes deviation from observed or boundary values,
- Nonlocal-consistency loss (): enforces agreement between learned memory field and its integral representation,
For fractional regimes:
These integrals are computed only inside via low-cost quadrature (e.g., Gauss–Legendre).
4. Operator Inversion and Solution Recovery
MINPO exploits the injectivity and analytic properties of the memory operator (with known kernel type), ensuring reconstruction via closed-form identities derived from Leibniz’ rule or fractional-calculus relations. Once (and for ) are learned, is recovered by differentiating and composing with learned inverse, without need for iterative inversion, history tracking, or explicit quadrature during inference.
This explicit treatment distinguishes MINPO from solvers that require repeated evaluation or accumulation of nonlocal integrals.
5. Empirical Evaluation Across Representative Nonlocal Problems
MINPO has been quantitatively benchmarked against classical spectral/fd solvers, A-PINN/fPINN, and their newly-introduced KAN analogues (A-PIKAN/fPIKAN).
Experiment I: Nonlinear Volterra IDE (1D, Forward & Inverse)
- Equation: ; analytic solution
- Forward ( known):
- MINPO-KAN (): ,
- A-PIKAN: ,
- A-PINN:
- MINPO-KAN reduces solution error by $35$– (vs. A-PIKAN), $55$– (vs. A-PINN)
- Inverse ( unknown, 10 measurements):
- MINPO-KAN: error –, substantially lower solution/operator errors than baselines
Experiment II: 3D Nested Nonlocal IDE
- Equation:
- Boundary constraint for is hard-wired into architecture
- MINPO-KAN: , (Table 3)
- 76.7% u-error reduction and 88.9% M-error reduction, faster vs. finite-difference baseline
- MINPO-MLP achieves up to acceleration
Experiment III: 1D Time-Fractional Diffusion
- Equation: , Dirichlet initial/boundary data
- MINPO computes
- For coarse , MINPO-KAN -error –, -error –, whereas fPIKAN/fPINN -error –, -error –
- MINPO yields uniformly improved accuracy as increases and smoother memory operator fields
Summary Performance Table
| Experiment | MINPO-KAN Accuracy | Competing Method Accuracy | Speedup |
|---|---|---|---|
| Volterra IDE | (A-PIKAN) | — | |
| 3D Nested IDE | |||
| Fractional PDE | -error $0.7$– | $1.8$– | — |
6. Generalization Properties and Prospective Applications
MINPO generalizes across IDE classes due to:
- Continuous enforcement of the governing physics; neither discretized quadrature nor collocation are required for nonlocal/fractional terms in the physics residual
- The nonlocal-consistency loss, computed via a single low-cost quadrature, acts as a regularizer independent of physics enforcement
- Explicit solution reconstruction, eliminating the need for cumulative history or discretization artifacts
- Flexibility to trade expressivity and parameter count between KANs and MLPs
Scalability
MINPO demonstrates robust performance in 1D, nested 3D, and fractional IDE settings. It avoids the curse of kernel dimensionality by having the physics residual term free of volumetric kernel summation, and achieves accuracy with orders-of-magnitude fewer degrees of freedom than finite-difference methods.
Potential Extensions
MINPO offers a framework suited to:
- Multi-term/distributed-order fractional PDEs
- Spatial fractional Laplacians, tempered Lévy kernels, peridynamics
- State-dependent or time-varying memory kernels (applications in hysteresis or adaptive systems)
- Integration with hybrid symbolic-neural kernel discovery via sparse regression
MINPO provides a unified framework to learn and solve IDEs of classical, fractional, spatially nonlocal, or mixed types by representing the nonlocal operator and its inverse as neural fields, enforcing the physics continuously, and reconstructing the solution via operator-theoretic identities. Across benchmark tasks, MINPO consistently outperforms classical solvers, A-PINN/fPINN, and their KAN variants, achieving up to an order-of-magnitude increase in operator accuracy and computational speedup of $5$– (Mostajeran et al., 19 Dec 2025).