Deep Onsager–Machlup Method (DOMM)
- Deep Onsager–Machlup Method (DOMM) is a mesh-free, DNN-parameterized variational approach to simulate irreversible soft-matter dynamics using the Onsager–Machlup principle.
- DOMM leverages residual neural networks and adaptive time-splitting to efficiently solve nonlinear, multi-scale, and multiphysics problems with an explicit thermodynamic structure.
- Demonstrations on diffusion, Cahn–Hilliard, and coupled Stokes dynamics validate DOMM’s accuracy and stability, achieving error thresholds and reducing training steps by 20–30%.
The Deep Onsager–Machlup Method (DOMM) is a computational framework for modeling soft matter dynamics. It synthesizes deep neural networks (DNNs) with the Onsager–Machlup variational principle (OMVP), enabling efficient, physics-informed solutions to nonlinear, multi-scale, and multiphysics problems characterized by multiple slow variables and dissipative processes. By replacing traditional hand-crafted ansatz functions with DNN-parameterized trial solutions, DOMM provides a mesh-free, variationally consistent strategy for simulating irreversible dynamics with explicit thermodynamic structure and high-order differential operators. DOMM can be viewed as an extension of the deep Ritz method, traditionally applied to elliptic and static problems, to dynamic settings governed by OMVP (Li et al., 2023).
1. Theoretical Foundations
DOMM is founded on the Onsager–Machlup variational principle, which provides a path-probability functional for the evolution of coarse-grained (slow) state variables . For dynamics over with rates and conjugate generalized forces , the Onsager–Machlup action is
where the OM functional
with and . In systems where the free energy defines conservative forces and additional non-conservative (active) forces may be present, this becomes
where is the mobility. Extremizing via the Euler–Lagrange equation yields the irreversible force-balance or gradient flow equations, e.g.,
2. DOMM Formulation
The distinctive feature of DOMM is the use of DNNs to parameterize the unknown dynamic field(s), typically denoted for order parameters, fluxes, or auxiliary quantities. The trial solution is expressed as where encompasses all network weights and biases. The architecture is generally a residual neural network (ResNet), described by
with as a smooth activation function to enable exact computation of high-order derivatives.
The loss functional minimized during training is
- : Onsager–Machlup (OM) action integral expressing thermodynamic irreversibility.
- : Quadratic penalty enforcing initial conditions.
- : Quadratic penalty enforcing boundary conditions (Dirichlet, Neumann, or periodic).
- : Penalty enforcing auxiliary constraints (mass conservation, incompressibility, etc.).
For order-parameter-conserving problems, is
where is the variationally optimal flux (e.g., in Cahn–Hilliard models).
3. Network Architecture, Collocation, and Training Procedure
The DOMM deploys fully connected ResNet architectures with depth –$6$ and width –$128$. The choice of cubic hyperbolic tangent activations ensures regularity of high-order derivatives, facilitating treatment of fourth-order operators.
Training is performed over a mesh-free set of collocation points sampled uniformly in . To improve computational scalability and leverage the “frequency principle” in DNN optimization, the solution domain in time is split into intervals of length , and training proceeds sequentially over This splitting reduces GPU memory requirements and preferentially optimizes smoother portions of the solution trajectory.
Mini-batching is employed by randomly sampling points from the collocation set per stochastic gradient descent (SGD) step. The Adam optimizer is used with learning rate –. Loss weights are empirically tuned to ensure comparable magnitudes among terms; initial and constraint penalties typically require larger weights. Convergence is monitored by both training loss and mean-square-error (MSE) at reference validation points, with rigorous criteria: for straightforward diffusion/Cahn–Hilliard problems and for coupled system dynamics.
4. Illustrative Applications
Three representative classes of soft matter PDEs were studied to validate DOMM (Li et al., 2023):
4.1 1D Particle Diffusion
For , trials with both step and Fourier-mode initial data and Neumann or Dirichlet BCs achieved in training steps (ResNet: , , tanh activation). Time-splitting with () and reduced data-per-step led to a $20$– reduction in steps to reach the same accuracy.
4.2 Two-Phase Cahn–Hilliard Dynamics
1D Cahn–Hilliard: For
with periodic boundaries and trigonometric initial data, DOMM achieved with , , and .
2D Droplet Coalescence: For
on with periodic BCs, initial data with two droplets, MSE was achieved using similar network settings; pointwise errors are near interfaces.
4.3 2D Stokes–Cahn–Hilliard Dynamics (Coupled Hydrodynamics)
PDEs:
Using , , , and weighted constraints, solution phase-fields and velocities matched finite-difference methods (FDM; ) with local errors and MSEMSE in steps.
5. Advantages, Limitations, and Comparison
Key features and outcomes demonstrated for DOMM (Li et al., 2023) include:
- Mesh-free, continuous representation, eliminating the need for spatial meshes and naturally accommodating complex geometries and dynamic interfaces.
- OM-based variational loss allows consistent coupling of multiple physical processes with thermodynamically derived weights—no manual term-scaling is required.
- High-order spatial and temporal derivatives (e.g., up to fourth order as in Cahn–Hilliard) are handled via automatic differentiation, circumventing the need for explicit finite-difference stencils.
- Time-splitting leverages the DNN “frequency principle” to focus on lower-frequency (smoother) solution components first, reducing long-time training costs.
- In comparison to the deep Ritz method (static problems) and PINNs (physics-informed neural networks), DOMM uniquely incorporates energy dissipation structures and most-probable-path criteria from OMVP, resulting in enhanced stability and fidelity for irreversible dynamics.
- Major limitations include scaling of training cost with input dimensionality and network size, sensitivity to relative weighting of loss terms, and challenges for very stiff or nonlinear kinetics, which may require adaptive sampling or curriculum-learning strategies.
6. Extensions and Future Directions
DOMM's variational strategy admits several extensions:
- Inverse Problems: Calibration of free-energy landscapes and mobility coefficients through minimization of with respect to experimental or trajectory data.
- Rare-Event Pathways: Identification of minimum-action (most-probable-path) solutions for nucleation, macromolecular dynamics, and related rare-event problems.
- Active Matter: Extension of OMVP to systems with non-conservative active forces , enabling learning of active transport coefficients and emergent flow characteristics.
- Multi-Scale Coupling: Embedding DOMM within hierarchical or domain-decomposition frameworks to address systems with disparate temporal or spatial scales, such as nematohydrodynamics or elasto-diffusion in gels.
DOMM provides a general, physics-informed, variationally consistent computational methodology for irreversible soft-matter dynamics, with rigorous enforcement of thermodynamic structure and verified predictive accuracy on canonical problems (Li et al., 2023).