Controlled Branching McKean-Vlasov Diffusion
- Controlled branching McKean–Vlasov diffusion is a framework that models particle dynamics with branching rates influenced by individual states and the overall population distribution.
- The system uses closed-loop control policies under Lipschitz constraints, leading to optimization problems characterized by nonlinear Fokker–Planck PDEs and HJB master equations.
- Explicit Riccati equations in linear–quadratic settings demonstrate tractable solutions for optimal feedback controls in complex stochastic environments.
A controlled branching McKean–Vlasov diffusion is a population system in which each particle’s stochastic dynamics, branching (birth-death) rates, and offspring distribution may depend both on the particle’s current position and the distribution (law) of the entire population. The system is subject to a closed-loop control policy that prescribes feedback controls stable under Lipschitz constraints. The control objective is formalized as an optimization problem on the space of probability measures, seeking to minimize (or maximize) an expected sum of running and terminal costs aggregated across all particles through their evolving measure. This framework yields a nonlinear Fokker–Planck PDE for the marginal law of the system under control and leads to an infinite-dimensional Hamilton–Jacobi–Bellman (HJB) master equation, with explicit solutions in certain linear–quadratic settings (Claisse et al., 29 Nov 2025).
1. Model Formulation
Controlled branching McKean–Vlasov diffusions generalize classical branching processes and mean-field diffusions by combining interactive branching particle genealogies with controlled stochastic flows. The population’s genealogy is encoded using the Ulam–Harris tree . At time , the alive set is ; the state of the full system is given by
where are particle positions. Mean–field interactions depend solely on the marginal empirical measure ,
Each particle evolves via a controlled stochastic differential equation (SDE), \begin{align*} dXk_s &= b\big(s,Xk_s,\mu_s,\alpha(s,Xk_s)\big)ds + \sigma\big(s,Xk_s,\mu_s,\alpha(s,Xk_s)\big)dWk_s, \end{align*} where are independent Brownian motions. Branching occurs at state-dependent rate , and after branching, the particle is replaced by offspring with probability , with all offspring locations coinciding with the parent.
The control is Lipschitz in and linear growth-bounded. The optimization objective uses a running cost and terminal cost . For initial measure at time , the objective is \begin{align*} J(t,\xi,\alpha) &= \mathbb{E}\left[\int_tT \sum_{k\in K_s} L(s,Xk_s,\mu_s,\alpha(s,Xk_s)) ds + \sum_{k\in K_T} g(Xk_T,\mu_T) \right] \ &= \int_tT \langle L(s,\cdot,\mu_s,\alpha(s,\cdot)), \mu_s\rangle ds + \langle g(\cdot, \mu_T), \mu_T \rangle, \end{align*} reflecting the deterministic evolution of .
2. Nonlinear Fokker–Planck Equation
For a given control , the marginal law evolves as a measure-valued solution to a nonlinear partial differential equation incorporating both diffusion and branching mechanisms. The branching increment is given by
For any test function ,
where
In density form, the evolution of satisfies: \begin{align*} \partial_s \mu_s + \nabla_x \cdot \Big(b(s,x,\mu_s,\alpha(s,x))\,\mu_s\Big) = \frac{1}{2} \sum_{i,j} \partial2_{x_i x_j}\left((\sigma\sigmaT)_{ij}(s, x, \mu_s, \alpha(s, x))\,\mu_s\right) \ + \pi(s, x, \mu_s, \alpha(s, x))\,\mu_s. \end{align*} This deterministic nonlinear equation encodes the measure evolution for mean-field control of the branching system.
3. Dynamic Programming and the Value Function
The optimal control problem is formulated for the value function defined on : \begin{align*} v(t, \nu) = \inf_{\alpha} \inf_{\xi: \mathcal{L}(\xi)=\nu} J(t, \xi, \alpha) = \inf_{\alpha}\left{ \int_tT \langle L(s, \cdot, \mu_s, \alpha(s, \cdot)), \mu_s \rangle ds + \langle g(\cdot, \mu_T), \mu_T \rangle \right}, \end{align*} where solves the Fokker–Planck equation from initial condition .
A dynamic programming principle (DPP) holds:
for any . This leverages the flow property of the controlled measure evolution and control concatenation.
4. HJB Master Equation in the Measure Space
Assuming the value function is classically regular in (using the Lions’ linear derivative with respect to measure), the infinite-dimensional HJB equation takes form: \begin{align*} \begin{cases} \partial_t v(t, m) + \inf_{a(\cdot)} \left{ \langle L(t, \cdot, m, a(\cdot)), m \rangle + \langle Ga_t v(t, m)(\cdot), m \rangle \right} = 0 \ v(T, m) = \langle g(\cdot, m), m \rangle \end{cases} \end{align*} where, for feedback law ,
Viscosity and classical solution concepts are distinguished by the regularity and test-function admissibility in this infinite-dimensional setting.
5. Verification Theorem and Optimal Feedback
Under the existence of a classical solution to the HJB master equation and for every a feedback attaining the infimum, the optimal closed-loop control is
With this feedback, the equality holds, certifying both optimality and attainment by the value function.
6. Explicit Solution in the Linear–Quadratic Setting
As a concrete example, specialization to the linear–quadratic case in one dimension (), with constant branching intensity and fixed offspring law , yields tractable Riccati equations:
- Drift and diffusion:
- Cost functionals:
where and .
- Value function Ansatz:
with .
Inserting into the HJB yields a coupled Riccati ODE system for . The optimal feedback is pointwise:
Direct substitution verifies as optimal and admissible, thereby solving the control problem in this regime and establishing explicit characterization of optimal closed-loop dynamics (Claisse et al., 29 Nov 2025).