Papers
Topics
Authors
Recent
Search
2000 character limit reached

PhICNet: Physics-Incorporated Conv RNN

Updated 20 January 2026
  • PhICNet is a hybrid framework integrating physics-based PDE models with deep learning to forecast spatio-temporal dynamics and identify hidden sources.
  • It employs a convolutional RNN core combined with a RED-Net for residual source prediction, achieving remarkable SNR improvements and high spatial correlation in source recovery.
  • The method is validated on canonical systems like heat diffusion, wave equations, and 2D Burgers’ flow, demonstrating robust forecasting performance and adaptability to noise and parameter shifts.

PhICNet (Physics-Incorporated Convolutional Recurrent Neural Network) is a hybrid modeling framework combining numerical physics-based models and deep learning for the joint tasks of spatio-temporal forecasting and source identification in dynamical systems governed by partial differential equations (PDEs) with unobservable, time-varying external sources. PhICNet formulates the underlying PDEs as convolutional RNN cells in a neural architecture, enabling data-driven learning of both the system evolution and hidden source terms. The framework is end-to-end trainable and yields state-of-the-art performance in forecasting and source recovery across a range of simulated physical systems (Saha et al., 2020).

1. Mathematical Formulation

PhICNet models a spatio-temporal field u(x,y,t)u(x,y,t) on a domain ΩR2\Omega \subset \mathbb{R}^2 governed by an inhomogeneous PDE of order nn in time: nutn=F(x,y,u,xu,yu,xxu,;θ)+v(x,y,t),(1)\frac{\partial^n u}{\partial t^n} = F(x,y,u,\partial_x u, \partial_y u, \partial_{xx} u, \ldots; \theta) + v(x,y,t), \tag{1} where FF is a known function (potentially nonlinear) of the field and its spatial derivatives, parameterized by unknown physical coefficients θ\theta (such as diffusivity α\alpha, wave speed cc, or viscosity β\beta). The unobservable source v(x,y,t)v(x,y,t) evolves via a separate kkth-order equation: kvtk=G(x,y,v).(2)\frac{\partial^k v}{\partial t^k} = G(x,y,v). \tag{2} Spatial discretization uses a regular X×YX \times Y grid with UtRX×YU_t \in \mathbb{R}^{X\times Y} and finite-difference approximations for temporal and spatial derivatives. Small (e.g., 3×33 \times 3) fixed convolutional kernels DjlD_{jl} implement spatial differencing, such as

D20=(000 121 000),    D02=(010 020 010).(3)D_{20} = \begin{pmatrix} 0 & 0 & 0 \ 1 & -2 & 1 \ 0 & 0 & 0 \end{pmatrix}, \;\; D_{02} = \begin{pmatrix} 0 & 1 & 0 \ 0 & -2 & 0 \ 0 & 1 & 0 \end{pmatrix}. \tag{3}

Temporal evolution is encoded in a convolutional RNN cell: the cell state CtRn×X×YC_t \in \mathbb{R}^{n\times X\times Y} stacks nn previous maps. The cell computes a homogeneous PDE prediction HtH_t by aggregating finite-difference temporal coefficients whcpw_{hc}^p and a (learnable) sum of weighted convolutional terms from the PDE's right-hand side: Ht=p=1nwhcpCtp+f(D10Ut,D01Ut,D20Ut,;θ).(6)H_t = \sum_{p=1}^n w_{hc}^p C_t^p + f(D_{10}*U_t, D_{01}*U_t, D_{20}*U_t, \ldots; \theta). \tag{6} The source term is estimated internally as a residual Vt=UtHt1V_t = U_t - H_{t-1} and buffered in CV,tRk×X×YC_{V,t} \in \mathbb{R}^{k \times X \times Y} for prediction of future sources via a RED-Net (residual encoder-decoder) network, yielding

V^t+1=p=1kwvcpCV,tp+g(CV,t),(9)\widehat V_{t+1} = \sum_{p=1}^k w_{vc}^p C_{V,t}^p + g(C_{V,t}), \tag{9}

with g()g(\cdot) parameterizing nonlinear source dynamics GG. The final prediction is the sum U^t+1=Ht+V^t+1\widehat U_{t+1} = H_t + \widehat V_{t+1}.

2. Network Architecture

PhICNet comprises two principal components:

  • PDE-RNN Core: Receives input UtU_t, maintains cell state buffer CtC_t for the last nn frames, and source buffer CV,tC_{V,t} for the last kk frames of estimated sources. All spatial operations leverage depth-wise 2D convolutions with fixed 3×33\times3 differential kernels. Physical parameters θ\theta are trainable scalars.
  • RED-Net for Source Dynamics: Adopts a classical residual encoder–decoder topology with symmetric skip connections. The encoder includes MM down-sampling blocks with two convolutional layers (typical dimensionality 64) and ReLU, mirrored by up-sampling blocks in the decoder. Output is added to the linear finite-difference source forecast.

The RED-Net refines the nonlinear evolution of the source term vv and allows end-to-end learning of complex time-varying perturbations.

3. Training and Loss Functions

Training sequences {U0,...,UT}\{U_0, ..., U_T\} are used to forecast {U^n+k,...,U^T}\{\widehat U_{n+k}, ..., \widehat U_T\} and estimate sources {V^n+k,...,V^T}\{\widehat V_{n+k}, ..., \widehat V_T\}. The composite loss function integrates:

  • Prediction Error:

Lpred=1T(n+k)+1t=n+kTUtU^t22.(11)\mathcal{L}_{\mathrm{pred}} = \frac{1}{T-(n+k)+1} \sum_{t=n+k}^T \|U_t - \widehat U_t\|_2^2. \tag{11}

  • Source-Prediction Consistency:

Lsrcpred=1T(n+k)+1t=n+kTVtV^t22,Vt=UtHt1.(12)\mathcal{L}_{\mathrm{src\,pred}} = \frac{1}{T-(n+k)+1} \sum_{t=n+k}^T \|V_t - \widehat V_t\|_2^2, \quad V_t = U_t - H_{t-1}. \tag{12}

  • Source Sparsity (optional):

Lsrcsparse=1T(n+k)+1t=n+kTV^t1.(13)\mathcal{L}_{\mathrm{src\,sparse}} = \frac{1}{T-(n+k)+1} \sum_{t=n+k}^T \|\widehat V_t\|_1. \tag{13}

Total loss: L=Lpred+Lsrcpred+λLsrcsparse.(14)\mathcal{L} = \mathcal{L}_{\mathrm{pred}} + \mathcal{L}_{\mathrm{src\,pred}} + \lambda\,\mathcal{L}_{\mathrm{src\,sparse}}. \tag{14} Optimization employs BPTT over all frames, with SGD or Adam at learning rate 10310^{-3}, batch size 4–8, and weight decay 10610^{-6}. Early stopping uses the validation loss.

4. Source Identification and Mechanisms

At each time step, PhICNet computes the homogeneous prediction Ht1H_{t-1}, calculates the empirical source residual Vt=UtHt1V_t = U_t - H_{t-1}, updates the source buffer, and predicts the next-step source V^t+1\widehat V_{t+1} via RED-Net and finite-difference aggregation. Training the source consistency loss encourages V^tVt\widehat V_t \approx V_t, and imposing Lsrcsparse\mathcal{L}_{\mathrm{src\,sparse}} enforces known sparsity. This mechanism enables direct recovery of the unobservable external sources driving the physical system, with internal source estimates accessible as by-products of forecasting.

5. Experimental Evaluation

PhICNet is evaluated on three canonical PDE systems, each discretized on 64×6464 \times 64 grids across a time horizon of T=200T=200 steps, with 20%20\% validation hold-out.

  • Heat Diffusion (n=1n=1):

vv is block-sparse and diffusive; 100 training, 50 test sequences.

  • Wave Equation (n=2n=2):

vv comprises two point oscillators at random locations; 300 training, 50 test sequences.

  • 2D Burgers’ Flow (n=1n=1, vector):

vv updates two Gaussian pressure zones traveling in circles; 300 training, 50 test sequences.

Boundary conditions: Dirichlet (heat, wave) and Neumann (Burgers).

Performance is measured by:

  • Forecast Accuracy: SNR in dB,

SNR(Ut,U^t)=20log10(Ut2UtU^t2).(15)\mathrm{SNR}(U_t, \widehat U_t) = 20 \log_{10} \left( \frac{\|U_t\|_2}{\|U_t-\widehat U_t\|_2} \right). \tag{15}

  • Source Identification: Spatial correlation ρ(Vttrue,Vtest)[1,1]\rho(V_t^{\mathrm{true}}, V_t^{\mathrm{est}}) \in [-1,1].

PhICNet is compared against three baselines: PDE-RNN + CNN, ConvLSTM, and RED-Net. The table below summarizes SNR and source correlation at t=150t=150 averaged over test sets (95% CI):

System Metric PDE-RNN+CNN ConvLSTM RED-Net PhICNet
Heat SNR@150 12 dB (±1) 18 dB (±2) 24 dB (±1.5) 30 dB (±1)
Corr(V)@150 0.40 (±0.05) 0.85 (±0.03)
Wave SNR@150 8 dB (±1) 14 dB (±1.8) 20 dB (±1.2) 26 dB (±1)
Corr(V)@150 0.30 (±0.06) 0.80 (±0.04)
Burgers SNR@150 10 dB (±1) 16 dB (±2) 22 dB (±1) 28 dB (±1)
Corr(V)@150 0.35 (±0.05) 0.78 (±0.05)

PhICNet achieves superior forecast longevity (SNR increases of 6–10 dB above baselines) and is uniquely effective in recovering unobservable source terms with spatial correlation ρ0.8\rho \approx 0.8–$0.9$.

6. Analysis: Hyperparameters, Robustness, and Adaptation

  • Sparsity Penalty λ\lambda:

Forecasting and source identification peak for 103λ10110^{-3} \leq \lambda \leq 10^{-1}. Excessive λ\lambda causes over-sparsification; insufficient values diminish identifiability.

  • Source Order kk:

Optimal kk matches the true source ODE order (e.g., k=2k=2 for wave experiments), whereas mismatch (k=1k=1 or k=3k=3) degrades performance.

  • Noise Robustness:

Adding Gaussian noise up to 10%10\% of data variance leads to SNR degradation of less than 2 dB.

  • Online Adaptation of Physical Parameters:

Physical coefficients θ\theta (e.g., α\alpha, cc) can be adapted online when discrepancies exceed a threshold, allowing model re-tuning while keeping RED-Net fixed. Empirical parameter trajectories closely follow ground-truth time-varying profiles.

7. Implementation Protocol

Spatial discretization is performed on [0,1]2[0,1]^2 to 64×6464 \times 64 grids (Δx=Δy=1/63\Delta x = \Delta y = 1/63), with timestep Δt\Delta t subject to the CFL condition. Differential kernels DjlD_{jl} are statically assigned; physical parameters θ\theta are randomly initialized. PyTorch is used for network implementation, buffers CtC_t and CV,tC_{V,t} are realized via FIFO queues. RED-Net depth is set to M=4M=4 with 64 channels and ReLU activations. Training is performed for 100–200 epochs with early stopping on validation loss.

This configuration permits direct reproduction of the forecasting and hidden source identification results across arbitrarily parameterized dynamical systems governed by PDEs of the form (F,G)(F,G) (Saha et al., 2020).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to PhICNet.