PhICNet: Physics-Incorporated Conv RNN
- PhICNet is a hybrid framework integrating physics-based PDE models with deep learning to forecast spatio-temporal dynamics and identify hidden sources.
- It employs a convolutional RNN core combined with a RED-Net for residual source prediction, achieving remarkable SNR improvements and high spatial correlation in source recovery.
- The method is validated on canonical systems like heat diffusion, wave equations, and 2D Burgers’ flow, demonstrating robust forecasting performance and adaptability to noise and parameter shifts.
PhICNet (Physics-Incorporated Convolutional Recurrent Neural Network) is a hybrid modeling framework combining numerical physics-based models and deep learning for the joint tasks of spatio-temporal forecasting and source identification in dynamical systems governed by partial differential equations (PDEs) with unobservable, time-varying external sources. PhICNet formulates the underlying PDEs as convolutional RNN cells in a neural architecture, enabling data-driven learning of both the system evolution and hidden source terms. The framework is end-to-end trainable and yields state-of-the-art performance in forecasting and source recovery across a range of simulated physical systems (Saha et al., 2020).
1. Mathematical Formulation
PhICNet models a spatio-temporal field on a domain governed by an inhomogeneous PDE of order in time: where is a known function (potentially nonlinear) of the field and its spatial derivatives, parameterized by unknown physical coefficients (such as diffusivity , wave speed , or viscosity ). The unobservable source evolves via a separate th-order equation: Spatial discretization uses a regular grid with and finite-difference approximations for temporal and spatial derivatives. Small (e.g., ) fixed convolutional kernels implement spatial differencing, such as
Temporal evolution is encoded in a convolutional RNN cell: the cell state stacks previous maps. The cell computes a homogeneous PDE prediction by aggregating finite-difference temporal coefficients and a (learnable) sum of weighted convolutional terms from the PDE's right-hand side: The source term is estimated internally as a residual and buffered in for prediction of future sources via a RED-Net (residual encoder-decoder) network, yielding
with parameterizing nonlinear source dynamics . The final prediction is the sum .
2. Network Architecture
PhICNet comprises two principal components:
- PDE-RNN Core: Receives input , maintains cell state buffer for the last frames, and source buffer for the last frames of estimated sources. All spatial operations leverage depth-wise 2D convolutions with fixed differential kernels. Physical parameters are trainable scalars.
- RED-Net for Source Dynamics: Adopts a classical residual encoder–decoder topology with symmetric skip connections. The encoder includes down-sampling blocks with two convolutional layers (typical dimensionality 64) and ReLU, mirrored by up-sampling blocks in the decoder. Output is added to the linear finite-difference source forecast.
The RED-Net refines the nonlinear evolution of the source term and allows end-to-end learning of complex time-varying perturbations.
3. Training and Loss Functions
Training sequences are used to forecast and estimate sources . The composite loss function integrates:
- Prediction Error:
- Source-Prediction Consistency:
- Source Sparsity (optional):
Total loss: Optimization employs BPTT over all frames, with SGD or Adam at learning rate , batch size 4–8, and weight decay . Early stopping uses the validation loss.
4. Source Identification and Mechanisms
At each time step, PhICNet computes the homogeneous prediction , calculates the empirical source residual , updates the source buffer, and predicts the next-step source via RED-Net and finite-difference aggregation. Training the source consistency loss encourages , and imposing enforces known sparsity. This mechanism enables direct recovery of the unobservable external sources driving the physical system, with internal source estimates accessible as by-products of forecasting.
5. Experimental Evaluation
PhICNet is evaluated on three canonical PDE systems, each discretized on grids across a time horizon of steps, with validation hold-out.
- Heat Diffusion ():
is block-sparse and diffusive; 100 training, 50 test sequences.
- Wave Equation ():
comprises two point oscillators at random locations; 300 training, 50 test sequences.
- 2D Burgers’ Flow (, vector):
updates two Gaussian pressure zones traveling in circles; 300 training, 50 test sequences.
Boundary conditions: Dirichlet (heat, wave) and Neumann (Burgers).
Performance is measured by:
- Forecast Accuracy: SNR in dB,
- Source Identification: Spatial correlation .
PhICNet is compared against three baselines: PDE-RNN + CNN, ConvLSTM, and RED-Net. The table below summarizes SNR and source correlation at averaged over test sets (95% CI):
| System | Metric | PDE-RNN+CNN | ConvLSTM | RED-Net | PhICNet |
|---|---|---|---|---|---|
| Heat | SNR@150 | 12 dB (±1) | 18 dB (±2) | 24 dB (±1.5) | 30 dB (±1) |
| Corr(V)@150 | 0.40 (±0.05) | — | — | 0.85 (±0.03) | |
| Wave | SNR@150 | 8 dB (±1) | 14 dB (±1.8) | 20 dB (±1.2) | 26 dB (±1) |
| Corr(V)@150 | 0.30 (±0.06) | — | — | 0.80 (±0.04) | |
| Burgers | SNR@150 | 10 dB (±1) | 16 dB (±2) | 22 dB (±1) | 28 dB (±1) |
| Corr(V)@150 | 0.35 (±0.05) | — | — | 0.78 (±0.05) |
PhICNet achieves superior forecast longevity (SNR increases of 6–10 dB above baselines) and is uniquely effective in recovering unobservable source terms with spatial correlation –$0.9$.
6. Analysis: Hyperparameters, Robustness, and Adaptation
- Sparsity Penalty :
Forecasting and source identification peak for . Excessive causes over-sparsification; insufficient values diminish identifiability.
- Source Order :
Optimal matches the true source ODE order (e.g., for wave experiments), whereas mismatch ( or ) degrades performance.
- Noise Robustness:
Adding Gaussian noise up to of data variance leads to SNR degradation of less than 2 dB.
- Online Adaptation of Physical Parameters:
Physical coefficients (e.g., , ) can be adapted online when discrepancies exceed a threshold, allowing model re-tuning while keeping RED-Net fixed. Empirical parameter trajectories closely follow ground-truth time-varying profiles.
7. Implementation Protocol
Spatial discretization is performed on to grids (), with timestep subject to the CFL condition. Differential kernels are statically assigned; physical parameters are randomly initialized. PyTorch is used for network implementation, buffers and are realized via FIFO queues. RED-Net depth is set to with 64 channels and ReLU activations. Training is performed for 100–200 epochs with early stopping on validation loss.
This configuration permits direct reproduction of the forecasting and hidden source identification results across arbitrarily parameterized dynamical systems governed by PDEs of the form (Saha et al., 2020).