Papers
Topics
Authors
Recent
2000 character limit reached

DUALFloodGNN: Physically-Informed Flood Modeling

Updated 4 January 2026
  • The paper demonstrates that DUALFloodGNN achieves improved predictive accuracy and efficiency by integrating explicit hydrodynamic variables and physics-based conservation laws.
  • DUALFloodGNN employs a unified message-passing mechanism and dynamic curriculum learning to enforce both global and local mass conservation for stable multi-step forecasts.
  • Experimental results show significant reductions in node-volume, edge-flow, and water-depth RMSE alongside orders-of-magnitude faster inference compared to traditional numerical solvers.

DUALFloodGNN is a physics-informed graph neural network architecture designed to enable rapid and physically consistent operational flood modeling on unstructured spatial domains. It extends standard Encode–Process–Decode GNN frameworks with explicit representations of hydrodynamic quantities and conservation laws, achieving improved predictive accuracy and computational efficiency compared to previous flood modeling approaches that leverage deep learning on spatial graphs (Acosta et al., 30 Dec 2025).

1. Model Architecture

DUALFloodGNN represents the computational domain as a directed graph G=(V,E)\mathcal{G}=(V,E), where each node i∈Vi\in V corresponds to a cell centroid in a 2D unstructured mesh, and each directed edge (i→j)∈E(i\to j)\in E represents adjacency between cells. Node features X∈R∣V∣×fvX\in \mathbb{R}^{|V|\times f_v} combine static attributes (cell area, elevation) and dynamic variables (water volume VtV^t, rainfall RR, upstream inflow QinQ_\text{in}, downstream outflow QoutQ_\text{out}) concatenated over pp previous time steps. Edge features E∈R∣E∣×feE\in \mathbb{R}^{|E|\times f_e} encode static properties (channel width, slope) and dynamic flow QtQ^t.

The model does not directly forecast future states Vt+1V^{t+1} and Qt+1Q^{t+1} but predicts their residual increments ΔV\Delta V and ΔQ\Delta Q, which are added to the latest state, improving stability in multi-step inference.

Input Construction: Xt=[XS ∥ XDt−p ∥ ⋯ ∥ XDt],Et=[ES ∥ EDt−p ∥ ⋯ ∥ EDt]\mathbf{X}^t = [X_S \,\Vert\, X_D^{t-p}\, \Vert\, \cdots \,\Vert\, X_D^t],\qquad \mathbf{E}^t = [E_S \,\Vert\, E_D^{t-p}\, \Vert\, \cdots \,\Vert\, E_D^t]

Encoder Mapping: H(0)=MLPnode(Xt),ε(0)=MLPedge(Et)H^{(0)} = \text{MLP}_\text{node}(X^t), \quad \varepsilon^{(0)} = \text{MLP}_\text{edge}(E^t)

Message Passing: For each layer ℓ=0,…,L−1\ell=0,\ldots,L-1, node and edge embeddings are jointly updated using messages: mji(ℓ)=MLPmsg([hi(ℓ) ∥ hj(ℓ) ∥ eij(ℓ)]),j∈N(i)m_{ji}^{(\ell)} = \text{MLP}_\text{msg}\left( [h_i^{(\ell)}\,\Vert\,h_j^{(\ell)}\,\Vert\, e_{ij}^{(\ell)} ] \right), \quad j\in \mathcal{N}(i)

hi(ℓ+1)=MLPupdate(∑j∈N(i)mji(ℓ))h_i^{(\ell+1)} = \text{MLP}_\text{update}\left( \sum_{j\in \mathcal{N}(i)} m_{ji}^{(\ell)} \right)

eij(â„“+1)=mji(â„“)e_{ij}^{(\ell+1)} = m_{ji}^{(\ell)}

A ReLU activation is applied at intermediate layers (except final), and biases are omitted in all MLPs to prevent spurious isolated-node volumes.

Decoding:

After LL process layers,

ΔVt+1=MLPdec,node(HL),ΔQt+1=MLPdec,edge(εL)\Delta V^{t+1} = \text{MLP}_\text{dec,node}(H^L), \qquad \Delta Q^{t+1} = \text{MLP}_\text{dec,edge}(\varepsilon^L)

2. Physics-Informed Loss Formulations

Training is governed by a composite loss function incorporating both standard prediction errors and physics-informed penalties that enforce mass conservation.

Prediction Loss (Node and Edge MSE): Lnode=1∣V∣∑i∈V(ΔVi‾−ΔVi)2,Ledge=1∣E∣∑k∈E(ΔQk‾−ΔQk)2\mathcal{L}_\text{node} = \frac{1}{|V|}\sum_{i\in V}\left( \overline{\Delta V_i} - \Delta V_i \right)^2,\quad \mathcal{L}_\text{edge} = \frac{1}{|E|}\sum_{k\in E}\left( \overline{\Delta Q_k} - \Delta Q_k\right)^2

Lpred=λ1 Lnode+λ2 Ledge\mathcal{L}_\text{pred} = \lambda_1\,\mathcal{L}_\text{node} + \lambda_2\,\mathcal{L}_\text{edge}

Global Mass Conservation: ∑iΔVit=(Qint−Qoutt)Δt+∑iRit\sum_{i} \Delta V_i^t = (Q_\text{in}^t - Q_\text{out}^t)\Delta t + \sum_{i} R_i^t

Lglobal=∣∑iΔVit−[(Qint−Qoutt)Δt+∑iRit]∣\mathcal{L}_\text{global} = \left\lvert \sum_{i}\Delta V_i^t - \left[(Q_\text{in}^t - Q_\text{out}^t)\Delta t + \sum_i R_i^t\right] \right\rvert

Local Mass Conservation (Nodewise):

Construct weighted adjacency IwI_w,

Iw=Iundir⊙[ReLU(Q) ∥ −ReLU(−Q)]I_w = I_\text{undir} \odot [\text{ReLU}(Q) \,\Vert\, -\text{ReLU}(-Q)]

Inflow and outflow per node ii: Qi+=∑j∈Nin(i)Iwj,i,Qi−=∑j∈Nout(i)Iwi,jQ_{i+} = \sum_{j\in\mathcal{N}_\text{in}(i)} I_{w_{j,i}},\quad Q_{i-} = \sum_{j\in\mathcal{N}_\text{out}(i)} I_{w_{i,j}} Local physical relation: ΔVit=(Qi+t−Qi−t)Δt+Rit\Delta V_i^t = (Q_{i+}^t - Q_{i-}^t)\Delta t + R_i^t Corresponding loss: Llocal=∑i∈V∣ΔVit−[(Qi+t−Qi−t)Δt+Rit]∣\mathcal{L}_\text{local} = \sum_{i\in V} \left\lvert \Delta V_i^t - \left[(Q_{i+}^t - Q_{i-}^t)\Delta t + R_i^t \right] \right\rvert

Total Physics-Informed Loss: Lphysics=λ3 Lglobal+λ4 Llocal\mathcal{L}_\text{physics} = \lambda_3\,\mathcal{L}_\text{global} + \lambda_4\,\mathcal{L}_\text{local}

Single-Step Loss: L=Lpred+Lphysics\mathcal{L} = \mathcal{L}_\text{pred} + \mathcal{L}_\text{physics}

3. Joint Prediction of Volume and Flow

DUALFloodGNN employs a unified message-passing mechanism whereby the shared messages mjim_{ji} update both node and edge latent states, thus enforcing intrinsic coupling between water volume (scalar-like) and flow (vector-like) variables. Overwriting of edge embeddings with the same message that updates node states enables information about transport and accumulation to coexist within a single latent space. Distinct decoders for ΔV\Delta V and ΔQ\Delta Q yield the final predictions for nodes and edges, respectively. This methodology obviates the computational expense of line-graph constructions and produces physically consistent multi-output inference in a single forward pass.

4. Multi-Step Training and Dynamic Curriculum

The model is optimized not only for single-step predictions but also for accurate multi-step autoregressive rollout. During training, for batch time-steps o=1,…,Oo=1,\dots,O, the model is unrolled into OO-step sequences, using its own predicted increments at each step for input construction. The rollout loss is the mean over the temporal sequence: Lrollout=1O∑o=1OL(ΔVo,ΔQo)\mathcal{L}_\text{rollout} = \frac{1}{O} \sum_{o=1}^{O} \mathcal{L}(\Delta V^o, \Delta Q^o)

Dynamic curriculum learning is employed: training starts with O=1O=1, and only after convergence at each rollout horizon is the sequence length incremented by CC (typically C=1C=1), at which point the learning rate decays by a factor γ\gamma. This staged progression stabilizes learning and mitigates error amplification in long-horizon autoregressive simulations.

5. Empirical Performance and Evaluation

On a Wollombi River test set comprising 1129 nodes, 2743 edges, 56 events, and using 14-fold cross-validation, DUALFloodGNN exhibits substantial quantitative improvements over alternative GNN architectures:

Metric DUALFloodGNN Best Comparator Relative Improvement
Node-volume RMSE ≈2.2×103\approx 2.2\times10^3 m³ GINE ≈3.4×103\approx 3.4\times10^3 m³ −35%-35\%
Edge-flow RMSE $25.9$ m³/s GraphSAGE edge: $44.0$ m³/s −41%-41\%
Water-depth RMSE $0.21$ m HydroGraphNet: $0.76$ m −72%-72\%
Flood CSI (0.05 m depth) $0.90$ HydroGraphNet: $0.69$; SWE-GNN: $0.47$
Inference time, 576 steps ∼4.1\sim 4.1 s HydroGraphNet: ∼3.3\sim 3.3 s Orders faster than full numerical solvers

Temporal RMSE plots confirm the persistence of low error across rollout, whereas competitors demonstrate more rapid error growth. Visual inundation maps show that DUALFloodGNN more closely tracks the ground truth, especially in the resolution of microscale lateral flows.

6. Implementation Details and Computational Scaling

The model is implemented in PyTorch and PyTorch-Geometric, with computation on NVIDIA V100 GPUs. Typical training uses L=4L=4 GNN layers, MLPs with width 128, and dropout rate 0.1. Loss weights (λ1,λ2,λ3,λ4)(\lambda_1,\lambda_2,\lambda_3,\lambda_4) are optimized by grid search; for example (1.0,1.0,0.5,1.0)(1.0,1.0,0.5,1.0). The curriculum step C=1C=1 and learning rate decay γ=0.8\gamma=0.8 are standard.

Training time is approximately 2 hours per cross-validation fold, with memory usage around 4 GB per V100 GPU. The method scales linearly with mesh size ∣V∣+∣E∣|V|+|E| and is demonstrated to handle up to 10510^5 cells in multi-GPU environments.

7. Limitations and Prospective Extensions

Although validated on flow-dominated events in a single catchment, DUALFloodGNN's generalization to rainfall-dominated floods, urban drainage systems, or coastal flooding scenarios—including significant wave effects—remains to be established. The use of absolute-value physics losses produces sub-gradients only at zero; investigation of alternative (e.g., Huber-type) residual penalties is warranted. Representation of complex, time-dependent boundary conditions such as moving stage hydrographs is not yet incorporated, and would require further architectural developments. Integration with uncertainty quantification frameworks, for example Bayesian GNNs or model ensembles, is necessary for operational deployment with reliable confidence intervals.

The principal contribution of DUALFloodGNN lies in its joint encoding of hydrodynamic volume and flow variables, its explicit enforcement of global and local conservation laws through loss function design, and its robust training strategy for multi-step inference. Open-source code and data-processing resources are provided to support reproduction and adaptation (Acosta et al., 30 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to DUALFloodGNN.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube