Papers
Topics
Authors
Recent
Search
2000 character limit reached

Controlled Branching McKean-Vlasov Diffusion

Updated 22 January 2026
  • Controlled branching McKean–Vlasov diffusion is a framework that models particle dynamics with branching rates influenced by individual states and the overall population distribution.
  • The system uses closed-loop control policies under Lipschitz constraints, leading to optimization problems characterized by nonlinear Fokker–Planck PDEs and HJB master equations.
  • Explicit Riccati equations in linear–quadratic settings demonstrate tractable solutions for optimal feedback controls in complex stochastic environments.

A controlled branching McKean–Vlasov diffusion is a population system in which each particle’s stochastic dynamics, branching (birth-death) rates, and offspring distribution may depend both on the particle’s current position and the distribution (law) of the entire population. The system is subject to a closed-loop control policy that prescribes feedback controls stable under Lipschitz constraints. The control objective is formalized as an optimization problem on the space of probability measures, seeking to minimize (or maximize) an expected sum of running and terminal costs aggregated across all particles through their evolving measure. This framework yields a nonlinear Fokker–Planck PDE for the marginal law of the system under control and leads to an infinite-dimensional Hamilton–Jacobi–Bellman (HJB) master equation, with explicit solutions in certain linear–quadratic settings (Claisse et al., 29 Nov 2025).

1. Model Formulation

Controlled branching McKean–Vlasov diffusions generalize classical branching processes and mean-field diffusions by combining interactive branching particle genealogies with controlled stochastic flows. The population’s genealogy is encoded using the Ulam–Harris tree KK. At time ss, the alive set is KsKK_s\subset K; the state of the full system is given by

Zs=kKsδ(k,Xsk)M(K×Rd),Z_s = \sum_{k\in K_s} \delta_{(k,X^k_s)} \in M(K\times\mathbb{R}^d),

where XskRdX^k_s\in\mathbb{R}^d are particle positions. Mean–field interactions depend solely on the marginal empirical measure μsM2(Rd)\mu_s \in M_2(\mathbb{R}^d),

μs(dx)=E[kKsδXsk(dx)].\mu_s(dx) = \mathbb{E}\left[\sum_{k\in K_s}\delta_{X^k_s}(dx)\right].

Each particle kk evolves via a controlled stochastic differential equation (SDE), \begin{align*} dXk_s &= b\big(s,Xk_s,\mu_s,\alpha(s,Xk_s)\big)ds + \sigma\big(s,Xk_s,\mu_s,\alpha(s,Xk_s)\big)dWk_s, \end{align*} where WkW^k are independent Brownian motions. Branching occurs at state-dependent rate γ(s,Xsk,μs,α(s,Xsk))\gamma(s,X^k_s,\mu_s,\alpha(s,X^k_s)), and after branching, the particle is replaced by \ell offspring with probability p(s,Xsk,μs,α(s,Xsk))p_\ell(s, X^k_s, \mu_s, \alpha(s, X^k_s)), with all offspring locations coinciding with the parent.

The control α:[0,T]×RdARn\alpha : [0,T]\times\mathbb{R}^d \to A\subset\mathbb{R}^n is Lipschitz in xx and linear growth-bounded. The optimization objective uses a running cost L:[0,T]×Rd×M2(Rd)×ARL:[0,T]\times\mathbb{R}^d\times M_2(\mathbb{R}^d)\times A \to\mathbb{R} and terminal cost g:Rd×M2(Rd)Rg:\mathbb{R}^d\times M_2(\mathbb{R}^d)\to\mathbb{R}. For initial measure ξ\xi at time tt, the objective is \begin{align*} J(t,\xi,\alpha) &= \mathbb{E}\left[\int_tT \sum_{k\in K_s} L(s,Xk_s,\mu_s,\alpha(s,Xk_s)) ds + \sum_{k\in K_T} g(Xk_T,\mu_T) \right] \ &= \int_tT \langle L(s,\cdot,\mu_s,\alpha(s,\cdot)), \mu_s\rangle ds + \langle g(\cdot, \mu_T), \mu_T \rangle, \end{align*} reflecting the deterministic evolution of μs\mu_s.

2. Nonlinear Fokker–Planck Equation

For a given control α\alpha, the marginal law μs\mu_s evolves as a measure-valued solution to a nonlinear partial differential equation incorporating both diffusion and branching mechanisms. The branching increment is given by

π(s,x,μ,a)=γ(s,x,μ,a)0(1)p(s,x,μ,a).\pi(s, x, \mu, a) = \gamma(s, x, \mu, a)\sum_{\ell\ge 0} (\ell-1)p_\ell(s, x, \mu, a).

For any test function φCb2(Rd)\varphi \in C^2_b(\mathbb{R}^d),

ddsμs,φ=μs,Lsαφ+π(s,,μs,α(s,))φ,\frac{d}{ds}\langle\mu_s, \varphi\rangle = \langle \mu_s, L^\alpha_s\varphi + \pi(s,\cdot, \mu_s, \alpha(s,\cdot))\,\varphi \rangle,

where

Lsαφ(x)=b(s,x,μs,α(s,x))φ(x)+12Tr[σσT(s,x,μs,α(s,x))2φ(x)].L^\alpha_s\varphi(x) = b(s, x, \mu_s, \alpha(s, x)) \cdot \nabla\varphi(x) + \tfrac12 \mathrm{Tr}\left[ \sigma \sigma^T (s, x, \mu_s, \alpha(s, x)) \nabla^2\varphi(x) \right].

In density form, the evolution of μs\mu_s satisfies: \begin{align*} \partial_s \mu_s + \nabla_x \cdot \Big(b(s,x,\mu_s,\alpha(s,x))\,\mu_s\Big) = \frac{1}{2} \sum_{i,j} \partial2_{x_i x_j}\left((\sigma\sigmaT)_{ij}(s, x, \mu_s, \alpha(s, x))\,\mu_s\right) \ + \pi(s, x, \mu_s, \alpha(s, x))\,\mu_s. \end{align*} This deterministic nonlinear equation encodes the measure evolution for mean-field control of the branching system.

3. Dynamic Programming and the Value Function

The optimal control problem is formulated for the value function vv defined on [0,T]×M2(Rd)[0,T]\times M_2(\mathbb{R}^d): \begin{align*} v(t, \nu) = \inf_{\alpha} \inf_{\xi: \mathcal{L}(\xi)=\nu} J(t, \xi, \alpha) = \inf_{\alpha}\left{ \int_tT \langle L(s, \cdot, \mu_s, \alpha(s, \cdot)), \mu_s \rangle ds + \langle g(\cdot, \mu_T), \mu_T \rangle \right}, \end{align*} where μs\mu_s solves the Fokker–Planck equation from initial condition ν\nu.

A dynamic programming principle (DPP) holds:

v(t,ν)=infα{tsL(u,,μu,α(u,)),μudu+v(s,μs)},v(t, \nu) = \inf_{\alpha} \left\{ \int_t^s \langle L(u, \cdot, \mu_u, \alpha(u, \cdot)), \mu_u \rangle du + v(s, \mu_s) \right\},

for any tsTt \leq s \leq T. This leverages the flow property of the controlled measure evolution and control concatenation.

4. HJB Master Equation in the Measure Space

Assuming the value function vv is classically regular in C1,2([0,T]×M2(Rd))C^{1,2}([0,T]\times M_2(\mathbb{R}^d)) (using the Lions’ linear derivative with respect to measure), the infinite-dimensional HJB equation takes form: \begin{align*} \begin{cases} \partial_t v(t, m) + \inf_{a(\cdot)} \left{ \langle L(t, \cdot, m, a(\cdot)), m \rangle + \langle Ga_t v(t, m)(\cdot), m \rangle \right} = 0 \ v(T, m) = \langle g(\cdot, m), m \rangle \end{cases} \end{align*} where, for feedback law a:RdAa: \mathbb{R}^d \to A,

Gtav(t,m)(x)=b(t,x,m,a(x))Dμv(t,m,x)+12Tr(σσTxDμv)(t,m,x)+γ0(1)pδμv(t,m,x).G^a_t v(t, m)(x) = b(t, x, m, a(x)) \cdot D_\mu v(t, m, x) + \frac12 \mathrm{Tr}\big(\sigma \sigma^T \partial_x D_\mu v\big)(t, m, x) + \gamma \sum_{\ell\ge 0} (\ell-1)p_\ell\,\delta_\mu v(t, m, x).

Viscosity and classical solution concepts are distinguished by the regularity and test-function admissibility in this infinite-dimensional setting.

5. Verification Theorem and Optimal Feedback

Under the existence of a classical solution uC1,2u \in C^{1,2} to the HJB master equation and for every (t,m)(t, m) a feedback a^(t,)\hat a(t, \cdot) attaining the infimum, the optimal closed-loop control is

αs(x)=a^(s,x,μst,m).\alpha^*_s(x) = \hat a(s, x, \mu^{t,m}_s).

With this feedback, the equality u(t,m)=v(t,m)u(t, m) = v(t, m) holds, certifying both optimality and attainment by the value function.

6. Explicit Solution in the Linear–Quadratic Setting

As a concrete example, specialization to the linear–quadratic case in one dimension (d=1d=1), with constant branching intensity γ\gamma and fixed offspring law pp_\ell, yields tractable Riccati equations:

  • Drift and diffusion:

b(s,x,m,a)=b1(s)x+b2(s)m(R)+b3(s)a,σ(s,x,m,a)=σ.b(s, x, m, a) = b_1(s)x + b_2(s)m(\mathbb{R}) + b_3(s)a, \qquad \sigma(s, x, m, a) = \sigma.

  • Cost functionals:

L(s,x,m,a)=L1(s)x2+L2(s)mˉ+L3(s)m1+L4(s)a2,g(x,m)=g1x2+g2mˉ+g3m1,L(s,x,m,a) = L_1(s)x^2 + L_2(s)\bar m + L_3(s)m_1 + L_4(s)a^2, \qquad g(x,m) = g_1 x^2 + g_2 \bar m + g_3 m_1,

where mˉ=m(R)\bar m = m(\mathbb{R}) and m1=xm(dx)m_1 = \int x\, m(dx).

  • Value function Ansatz:

v(t,m)=Λ(t)m2+Γ1(t)mˉ+Γ2(t)mˉm1+Γ3(t)mˉ2+Γ4(t)mˉ3,v(t,m) = \Lambda(t)m_2 + \Gamma_1(t)\bar m + \Gamma_2(t)\bar m m_1 + \Gamma_3(t)\bar m^2 + \Gamma_4(t)\bar m^3,

with m2=x2m(dx)m_2 = \int x^2\, m(dx).

Inserting into the HJB yields a coupled Riccati ODE system for (Λ,Γ1,Γ2,Γ3,Γ4)(\Lambda, \Gamma_1, \Gamma_2, \Gamma_3, \Gamma_4). The optimal feedback is pointwise:

α(t,x,m)=12L4(t)[2b3(t)Λ(t)x+b3(t)Γ2(t)mˉ].\alpha^*(t, x, m) = -\frac{1}{2L_4(t)} \left[ 2b_3(t)\Lambda(t)x + b_3(t)\Gamma_2(t)\bar m \right].

Direct substitution verifies (v,α)(v, \alpha^*) as optimal and admissible, thereby solving the control problem in this regime and establishing explicit characterization of optimal closed-loop dynamics (Claisse et al., 29 Nov 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Controlled Branching McKean-Vlasov Diffusion.