Forward PINN for Mesh-Free PDE Simulation
- Forward PINN is a physics-informed neural network that solves forward PDEs by embedding governing equations into its residual-based loss function.
- It employs deep neural networks to approximate fields in advection–dispersion problems, achieving relative L2 errors below 1% in benchmark tests.
- The mesh-free approach circumvents classical discretization challenges, enabling robust performance in irregular domains and high-dimensional settings.
A Forward PINN (“Physics-Informed Neural Network”) is a machine learning model designed to solve forward problems in partial differential equations (PDEs) by encoding physics directly into the loss function. In these applications, the PINN is trained to approximate the solution of a known PDE given initial and boundary conditions, without relying on discretized meshes or explicit numerical stencils. The method replaces classical mesh-based solvers with deep neural networks, which can be trained to minimize residuals of the governing equations and satisfy constraints, offering flexibility in irregular domains and high-dimensional problems.
1. Mathematical Formulation of the Forward PINN
The forward PINN method addresses the transient advection–dispersion equation (ADE) for solute transport in porous media: subject to initial and boundary conditions: Here, is the scalar concentration, is known velocity, is the (possibly anisotropic) dispersion tensor, and is the computational domain with Dirichlet and Neumann boundaries. The Peclét number () often characterizes the ratio of advection to dispersion.
2. Neural Network Architecture for Physics-Imposed Solution
In the forward ADE/Darcy context, separate deep neural networks (DNNs) are constructed for each field:
- Conductivity DNN: , typically layers, activations, interpolates known values for heterogeneous hydraulic conductivity.
- Hydraulic Head DNN: , layers, activations, solves .
- Concentration DNN: , layers, activations, solves the transient ADE.
Network inputs are the spatial and/or temporal coordinates; outputs are scalar field predictions. Weight initialization employs Xavier schemes to facilitate training stability.
3. Residual-based PINN Loss Construction
The core of the forward PINN is the residual-driven loss, integrating physics constraints into neural optimization: These are translated into a total mean-squared loss: where is typical. Experimental settings often include –$200,000$ collocation points in space-time, , and for 1D problems, with higher numbers in 2D.
Optional terms can be added for available measurement data:
4. Training Algorithms and Hyperparameter Choices
The forward PINN is typically optimized in two stages:
- Adam optimizer: 20,000 iterations, learning rate , mini-batch size 500.
- L-BFGS-B optimizer: full-batch, until convergence, strict tolerance .
Weights in the loss function are set by physical scaling; works robustly in forward ADEs. Collocation sampling is uniform or quasi-random (e.g., Sobol/Latin-Hypercube) to ensure adequate domain coverage.
5. Quantitative Benchmarks and Comparison to Classical Methods
The primary figure of merit is the relative error over a test grid: Forward PINN results on benchmark ADEs include:
- 1D, Pe=62.8:
- 1D, Pe=628: max pointwise error, sharp gradient recovery requires more network capacity
- 2D, Pe=50–200:
The method remains accurate ( relative error) up to Pe=200, outperforming conventional Galerkin FEM and stabilized SUPG methods, particularly at large Pe where mesh-aligned methods suffer crosswind oscillations or tuning instability.
6. Advantages, Scalability, and Domain-specific Insights
PINN's main strengths for forward problems are:
- Discretization-free: Avoids meshes, enabling easy application to irregular or high-dimensional domains.
- Parameter-free stability: Does not require stabilization tuning, as with SUPG.
- Robustness to mesh orientation: No crosswind artifacts.
- Flexible enforcement of boundary/initial conditions.
- Easy inclusion of measurement data.
Computational cost scales with the number of dimensions and problem stiffness (Pe); GPU acceleration and adaptive residual sampling can ameliorate scaling. Training typically converges in minutes to hours (GPU), and cost is competitive with mesh-refined PDE solvers for 2D ADEs.
7. Practical Summary and Outlook
The forward PINN framework enforces a mesh-free, residual-based loss incorporating the governing PDE and constraints. Standard architectures (4–5 layers, 30–60 neurons) and a two-stage Adam → L-BFGS schedule suffice for robust training on challenging advection–dispersion systems. Relative errors below are achieved for Pe up to 200, with marked advantages over FEM in terms of stability and flexibility for complex geometries. This approach thus provides a powerful meshless alternative for forward PDE simulations, especially as computational resources and parallelism grow. Applications are well supported in both single-physics and coupled Darcy/ADE systems (He et al., 2020).