Papers
Topics
Authors
Recent
2000 character limit reached

Memory-Informed Neural Pseudo-Operator (MINPO)

Updated 23 December 2025
  • MINPO is a mesh-free neural framework that models nonlocal spatiotemporal dynamics and fractional PDEs via learned memory operators and inverse reconstructions.
  • It integrates MLP and Kolmogorov-Arnold networks as encoders with a nonlocal-consistency loss to enforce coherence between the learned operator and the reconstructed dynamics.
  • Benchmark experiments show MINPO achieves significant error reductions and speedups compared to classical solvers and existing PINN approaches.

The Memory-Informed Neural Pseudo-Operator (MINPO) is a mesh-free neural framework designed to model and resolve nonlocal spatiotemporal dynamics described by integro-differential equations (IDEs). It encapsulates both the memory operator and its inverse through neural networks, enabling explicit reconstruction of unknown solution fields. The architecture leverages either multilayer perceptrons (MLPs) or Kolmogorov-Arnold Networks (KANs) as encoders and introduces a nonlocal consistency loss to enforce coherence between the learned operator field and the reconstructed dynamics. MINPO is applicable to systems governed by diverse nonlocal kernel structures, including fractional partial differential equations (PDEs), and demonstrates robust generalization across problem classes and kernel dimensionalities (Mostajeran et al., 19 Dec 2025).

1. Mathematical Formulation of Nonlocal Dynamics

MINPO targets the general form of an IDE embodying spatial nonlocality, temporal memory, and fractional dynamics. The governing equation in space-time coordinates ξ=(x,t)Ω×(0,T)\xi=(x,t)\in\Omega\times(0,T) is:

Tα[u](ξ)=N[u](ξ)+M[u](ξ)+S(ξ)T_\alpha[u](\xi) = N[u](\xi) + M[u](\xi) + S(\xi)

  • N[u](ξ)N[u](\xi): local (possibly nonlinear) differential operator
  • TαT_\alpha: composite time-evolution operator, including the first-order and fractional Caputo derivative,

Tα[u](ξ)=λ1ut(ξ)+λαCDtαu(ξ),0<α<1T_\alpha[u](\xi) = \lambda_1 \frac{\partial u}{\partial t}(\xi) + \lambda_\alpha \, {}^C D_t^\alpha u(\xi), \qquad 0<\alpha < 1

where CDtα{}^C D_t^\alpha is the Caputo derivative,

CDtαu(x,t)=1Γ(1α)0tu(x,τ)τ(tτ)α dτ{}^C D_t^\alpha u(x,t) = \frac{1}{\Gamma(1-\alpha)} \int_0^t \frac{\partial u(x,\tau)}{\partial \tau} (t-\tau)^{-\alpha}~d\tau

  • M[u](ξ)M[u](\xi): unified nonlocal memory operator, for general kernel KK,

M[u](ξ):=ΩξK(ξ,η;u(η))dηM[u](\xi) := \int_{\Omega_{\xi^-}} K(\xi,\eta;u(\eta))\,d\eta

where Ωξ\Omega_{\xi^-} denotes the causal past of ξ\xi.

The inverse of the Caputo derivative is the Riemann–Liouville fractional integral:

Itα[f](t)=1Γ(α)0t(tτ)α1f(τ)dτI_t^\alpha[f](t) = \frac{1}{\Gamma(\alpha)} \int_0^t (t-\tau)^{\alpha - 1} f(\tau)\,d\tau

Many classical and fractional nonlocal PDEs are subsumed as special cases of this general structure.

2. Neural Architecture and Operator Representation

MINPO learns the memory operator and its inverse via two neural networks:

  • Memory encoder Mθ(ξ)M[u](ξ)M_\theta(\xi) \approx M[u](\xi)
  • Inverse-memory encoder Jϕ(ξ)Itα[CDtαu](ξ)J_\phi(\xi) \approx I_t^\alpha[{}^C D_t^\alpha u](\xi) (for fractional dynamics)

Encoder Choices

  • MLP encoder: three hidden layers, 33 neurons per layer with tanh\tanh activation
  • KAN encoder: three hidden layers, 15 neurons, each neuron is a tunable Chebyshev-polynomial map (degree k=4k=4), with tanh\tanh nonlinearities between layers

KANs simultaneously learn linear weights and coefficients of univariate Chebyshev activation maps, yielding enhanced parameter efficiency for nonlocal operator learning.

Explicit Solution Reconstruction

Given MθM_\theta (and JϕJ_\phi in the fractional case), MINPO reconstructs the solution uu via an explicit differentiable ansatz:

uΘ(ξ)=u0(x)+γ0aγ(ξ)ξγMθ(ξ)+V~(ξ;Θ)u_\Theta(\xi) = u_0(x) + \sum_{|\gamma|\geq 0} a_\gamma(\xi)\,\nabla_\xi^\gamma M_\theta(\xi) + \tilde{V}(\xi;\Theta)

  • γ\gamma indexes spatial/time derivatives ξγ\nabla_\xi^\gamma
  • Coefficients aγ(ξ)a_\gamma(\xi) arise from applying Leibniz' rule (or corresponding identities for fractional inversion)
  • V~(ξ;Θ)\tilde{V}(\xi;\Theta) enforces initial traces or fractional inverse requirements

Common reduction cases:

Experiment Solution Reconstruction
Volterra IDE (Exp I) u(t)=tMθ(t)+Mθ(t)u(t) = \partial_t M_\theta(t) + M_\theta(t)
3D nested IDE (Exp II) u(x1,x2,t)=t,x1,x23Mθ+x1,x22Mθu(x_1,x_2,t) = \partial^3_{t,x_1,x_2} M_\theta + \partial^2_{x_1,x_2} M_\theta
Fractional PDE (Exp III) u(x,t)=u(x,0)+Jϕ(x,t)u(x,t) = u(x,0) + J_\phi(x,t)

3. Loss Functions and Training Procedure

All neural parameters Θ=(θ,ϕ)\Theta = (\theta, \phi) are jointly optimized using a composite loss function:

L(Θ)=λIDELIDE+λdataLdata+λMLML(\Theta) = \lambda_{\text{IDE}} L_{\text{IDE}} + \lambda_{\text{data}} L_{\text{data}} + \lambda_M L_M

  • Physics-residual loss (LIDEL_{\text{IDE}}): measures MSE residual over collocation points,

LIDE=MSE[Tα[uΘ]N[uΘ]MθS]L_{\text{IDE}} = \text{MSE} \left[ T_\alpha[u_\Theta] - N[u_\Theta] - M_\theta - S \right]

Crucially, no kernel or fractional term is discretized in this residual; all nonlocality resides in MθM_\theta and JϕJ_\phi.

  • Data-fidelity loss (LdataL_{\text{data}}): penalizes deviation from observed or boundary values,

Ldata=MSE[uΘuobs]L_{\text{data}} = \text{MSE}[u_\Theta - u_{\text{obs}}]

  • Nonlocal-consistency loss (LML_M): enforces agreement between learned memory field and its integral representation,

LM=MSE[Mθ(ξi)ΩξiK(ξi,η;uΘ(η))dη]L_M = \text{MSE} \left[ M_\theta(\xi_i) - \int_{\Omega_{\xi_i^-}} K(\xi_i, \eta; u_\Theta(\eta)) d\eta \right]

For fractional regimes:

LM=MSE[Mθ(x,t)1Γ(1α)0tτuΘ(x,τ)(tτ)αdτ]L_M = \text{MSE} \left[ M_\theta(x, t) - \frac{1}{\Gamma(1-\alpha)} \int_0^t \partial_\tau u_\Theta(x, \tau) (t - \tau)^{-\alpha} d\tau \right]

These integrals are computed only inside LML_M via low-cost quadrature (e.g., Gauss–Legendre).

4. Operator Inversion and Solution Recovery

MINPO exploits the injectivity and analytic properties of the memory operator M[u]M[u] (with known kernel type), ensuring reconstruction via closed-form identities derived from Leibniz’ rule or fractional-calculus relations. Once MθM_\theta (and JϕJ_\phi for α<1\alpha < 1) are learned, uΘu_\Theta is recovered by differentiating MθM_\theta and composing with learned inverse, without need for iterative inversion, history tracking, or explicit quadrature during inference.

This explicit treatment distinguishes MINPO from solvers that require repeated evaluation or accumulation of nonlocal integrals.

5. Empirical Evaluation Across Representative Nonlocal Problems

MINPO has been quantitatively benchmarked against classical spectral/fd solvers, A-PINN/fPINN, and their newly-introduced KAN analogues (A-PIKAN/fPIKAN).

Experiment I: Nonlinear Volterra IDE (1D, Forward & Inverse)

  • Equation: dudt+u=κ0teτtu(τ)dτ\displaystyle \frac{du}{dt} + u = \kappa \int_0^t e^{\tau-t} u(\tau) d\tau; analytic solution u(t)=etcosh(κt)u(t)=e^{-t}\cosh(\sqrt{\kappa}\, t)
  • Forward (κ\kappa known): NIDE=2400N_{\text{IDE}}=2400
    • MINPO-KAN (κ=1\kappa=1): E(u)1.49×104E(u)\approx1.49\times10^{-4}, E(M)1.89×104E(M)\approx1.89\times10^{-4}
    • A-PIKAN: 3.27×104/2.85×1043.27\times10^{-4}/2.85\times10^{-4},
    • A-PINN: 4.28×104/2.22×1044.28\times10^{-4}/2.22\times10^{-4}
    • MINPO-KAN reduces solution error by $35$–45%45\% (vs. A-PIKAN), $55$–65%65\% (vs. A-PINN)
  • Inverse (u,κu, \kappa unknown, 10 measurements):
    • MINPO-KAN: κ\kappa error O(105O(10^{-5}104)10^{-4}), substantially lower solution/operator errors than baselines

Experiment II: 3D Nested Nonlocal IDE

  • Equation: (t+x1+x2)u=u+0x20x10teτtu(y1,y2,τ)dτdy1dy2+f(ξ)(\partial_t+\partial_{x_1}+\partial_{x_2})u = u + \int_0^{x_2}\int_0^{x_1}\int_0^t e^{\tau-t} u(y_1,y_2,\tau) d\tau dy_1 dy_2 + f(\xi)
  • Boundary constraint for MM is hard-wired into architecture
  • MINPO-KAN: E(u)=4.49×103E(u)=4.49\times10^{-3}, E(M)=1.46×103E(M)=1.46\times10^{-3} (Table 3)
    • 76.7% u-error reduction and 88.9% M-error reduction, 4.7×4.7\times faster vs. finite-difference baseline
    • MINPO-MLP achieves up to 15.3×15.3\times acceleration

Experiment III: 1D Time-Fractional Diffusion

  • Equation: CDtαu(x,t)=x2u(x,t)+S(x,t){}^C D_t^\alpha u(x,t) = \partial_x^2 u(x,t) + S(x,t), Dirichlet initial/boundary data
  • MINPO computes u(x,t)=u(x,0)+Jϕ(x,t)u(x,t)=u(x,0)+J_\phi(x,t)
  • For coarse Nt=10N_t=10, MINPO-KAN uu-error 0.7\approx0.71.8%1.8\%, MM-error 2\approx26%6\%, whereas fPIKAN/fPINN uu-error 1.8\approx1.83.5%3.5\%, MM-error 10\approx1018%18\%
  • MINPO yields uniformly improved accuracy as NtN_t increases and smoother memory operator fields

Summary Performance Table

Experiment MINPO-KAN Accuracy Competing Method Accuracy Speedup
Volterra IDE E(u)1.49×104E(u)\approx1.49\times10^{-4} 3.27×1043.27\times10^{-4} (A-PIKAN)
3D Nested IDE E(u)=4.49×103E(u)=4.49\times10^{-3} 8.98×1038.98\times10^{-3} 4.7×4.7\times
Fractional PDE uu-error $0.7$–1.8%1.8\% $1.8$–3.5%3.5\%

6. Generalization Properties and Prospective Applications

MINPO generalizes across IDE classes due to:

  • Continuous enforcement of the governing physics; neither discretized quadrature nor collocation are required for nonlocal/fractional terms in the physics residual
  • The nonlocal-consistency loss, computed via a single low-cost quadrature, acts as a regularizer independent of physics enforcement
  • Explicit solution reconstruction, eliminating the need for cumulative history or discretization artifacts
  • Flexibility to trade expressivity and parameter count between KANs and MLPs

Scalability

MINPO demonstrates robust performance in 1D, nested 3D, and fractional IDE settings. It avoids the curse of kernel dimensionality by having the physics residual term free of volumetric kernel summation, and achieves accuracy with orders-of-magnitude fewer degrees of freedom than finite-difference methods.

Potential Extensions

MINPO offers a framework suited to:

  • Multi-term/distributed-order fractional PDEs
  • Spatial fractional Laplacians, tempered Lévy kernels, peridynamics
  • State-dependent or time-varying memory kernels (applications in hysteresis or adaptive systems)
  • Integration with hybrid symbolic-neural kernel discovery via sparse regression

MINPO provides a unified framework to learn and solve IDEs of classical, fractional, spatially nonlocal, or mixed types by representing the nonlocal operator and its inverse as neural fields, enforcing the physics continuously, and reconstructing the solution via operator-theoretic identities. Across benchmark tasks, MINPO consistently outperforms classical solvers, A-PINN/fPINN, and their KAN variants, achieving up to an order-of-magnitude increase in operator accuracy and computational speedup of $5$–15×15\times (Mostajeran et al., 19 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Memory-Informed Neural Pseudo-Operator (MINPO).