Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 34 tok/s Pro
GPT-5 High 36 tok/s Pro
GPT-4o 102 tok/s Pro
Kimi K2 195 tok/s Pro
GPT OSS 120B 433 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Sinkhorn Flow in Optimal Transport

Updated 16 October 2025
  • Sinkhorn Flow is the continuum limit of entropy-regularized Sinkhorn iterations, producing an absolutely continuous path in the 2-Wasserstein space.
  • It is characterized by a mirror gradient flow of the KL divergence with a non-standard velocity field derived from the Brenier map, linking optimal transport with parabolic Monge–Ampère dynamics.
  • The flow exhibits exponential convergence under logarithmic Sobolev conditions and is represented via a McKean–Vlasov diffusion, underpinning its applications in generative modeling and robust optimal transport.

Sinkhorn Flow is the continuum limit of the regularized Sinkhorn algorithm (or iterative proportional fitting procedure, IPFP) for entropy-regularized optimal transport, realized as the number of iterations grows while the entropic penalty vanishes with a precise scaling. In this limit, the alternating projections of the classical Sinkhorn algorithm converge to an absolutely continuous path in the 2-Wasserstein space, whose evolution is governed by a Wasserstein mirror gradient flow of the Kullback–Leibler (KL) divergence. This continuous-time dynamics can be described via both a continuity equation with a non-standard (mirror) velocity field and an associated parabolic Monge–Ampère equation for the dual Brenier potential, with convergence rates, well-posedness, and stochastic process representations under robust analytic conditions.

1. Definition and Conceptualization

The Sinkhorn flow is defined as the scaling limit of the Sinkhorn (IPFP) iterations for the entropically regularized optimal transport problem. Let the regularization parameter ε tend to zero, with the number of iterations proportional to 1/ε. Consider two probability densities, e{–f} (source) and e{–g} (target), on ℝd. When solving the entropy-regularized OT problem with the Sinkhorn algorithm, the sequence of marginal iterates in one coordinate converges, in the large-iteration and small-ε regime, to a continuous path (ρₜ)_t in the 2-Wasserstein space.

This path solves the continuity equation

tρt+div(ρtvt)=0\partial_t \rho_t + \mathrm{div}(\rho_t v_t) = 0

where the velocity field v_t is not the usual Wasserstein gradient of KL(ρ_t‖e{–f}) but a mirror gradient defined relative to the Brenier map transporting ρ_t to e{–g}. More precisely, the flow is the Wasserstein mirror gradient flow of the KL divergence, with mirror map U(ρ) = ½ W_22(ρ, e{–g}) and objective F(ρ) = KL(ρ‖e{–f}).

2. Mathematical Formulation

There are equivalent characterizations of the limiting dynamics:

  • Parabolic Monge–Ampère (PMA) Equation. The potential u_t evolves according to

tut(x)=f(x)g(x(ut))+det(x(ut)x),\partial_t u_t(x) = f(x) - g(x^{(u_t)}) + \det\left(\frac{\partial x^{(u_t)}}{\partial x}\right),

where x(ut)=ut(x)x^{(u_t)} = \nabla u_t(x) is the Brenier map.

  • Continuity Equation with Mirror Velocity. The marginal ρ_t is the pushforward (ut)#1eg(\nabla u_t)_{\#}^{-1} e^{–g} and solves

tρt+div(ρtvt)=0,vt(x)=(2ut(x))1[f(x)+logρt(x)].\partial_t \rho_t + \mathrm{div}(\rho_t v_t) = 0, \qquad v_t(x) = - (\nabla^2 u_t(x))^{-1} \nabla \left[f(x) + \log \rho_t(x)\right].

  • Mirror Gradient Flow Representation. The evolution is

t(ut)=WKL(ρtef),\partial_t (\nabla u_t) = \nabla_{W} KL(\rho_t \| e^{-f}),

i.e. the time derivative of the map matches the Wasserstein gradient of KL(ρ_t‖e{–f}).

The metric derivative of (ρ_t)_t with respect to the linearized optimal transport (LOT) distance is exactly the L2(ρt)L^2(\rho_t)-norm of v_t.

3. Convergence and Regularity Conditions

Rigorous convergence of discrete Sinkhorn iterates to Sinkhorn flow requires several regularity and convexity conditions:

  • The initial potential u_0, cost functions f, g are smooth with uniformly bounded derivatives up to order 3 or 4.
  • The Hessian matrices 2ut(x)\nabla^2 u_t(x) are uniformly positive definite and bounded on compact time intervals.
  • f and g satisfy a logarithmic Sobolev inequality (LSI), which is also used to establish exponential convergence to equilibrium.

Under these assumptions, the error in Wasserstein distance between the marginals produced by the Sinkhorn algorithm and the continuum Sinkhorn flow is O(ε)O(\varepsilon) (or o(ε)o(\varepsilon)) as ε→0.

4. Connection to Optimal Transport and Mirror Flows

Classical optimal transport uses the 2-Wasserstein metric, and its gradient flow for KL divergence is the standard Fokker–Planck equation. The Sinkhorn flow, by contrast, is a mirror gradient flow: the "mirror" is U(ρ)=½W22(ρ,eg)U(\rho) = ½ W_2^2(\rho, e^{–g}) and the objective is F(ρ)=KL(ρef)F(\rho) = KL(\rho \| e^{-f}). The gradient is taken in the mirror coordinates defined by the Brenier map. The velocity field vtv_t is not the Wasserstein gradient of the KL, but its pullback via the mirror transformation, and the metric speed is given by the L2L^2-norm of vtv_t with respect to a linearized optimal transport distance.

This structure allows the theory to connect the mirror descent interpretation of Sinkhorn (in the discrete setting) with the continuous-time flow, and, via the PMA equation, with the nonlinear geometry of optimal transport.

5. Parabolic Monge–Ampère Equation and Potential Representation

The Sinkhorn potential—the logarithmic scaling factor in the Sinkhorn algorithm—converges to the solution u_t(x) of the PMA equation:

tut(x)=f(x)g(x(ut))+det(x(ut)x)\partial_t u_t(x) = f(x) - g(x^{(u_t)}) + \det\left(\frac{\partial x^{(u_t)}}{\partial x}\right)

This gives a "potential" description of the entire flow, so that all the dynamics can be described as an evolution in the dual Brenier potential which generates the sequence of marginal distributions via the Brenier map.

The PMA equation also reinforces the interpretation of Sinkhorn flow as a mirror gradient flow: the evolution of u_t is the mirror descent of KL in the geometry generated by the squared Wasserstein distance to ege^{-g}.

6. Exponential Convergence and Logarithmic Sobolev Inequalities

If the target density efe^{-f} satisfies a logarithmic Sobolev inequality with constant λ>0\lambda > 0, then it is proved that the KL divergence decays exponentially along the Sinkhorn flow:

KL(ρtef)KL(ρ0ef)exp(2λ0th(s)ds)KL(\rho_t \| e^{-f}) \leq KL(\rho_0 \| e^{-f}) \exp\left(-2\lambda \int_0^t h(s) ds\right)

Here, h(s)h(s) is a time-dependent uniform lower bound on the metric induced by the Hessian of u_t. Exponential convergence in the Wasserstein distance can also be deduced via a suitable HWI inequality. This result parallels the exponential decay of entropy in classical Wasserstein gradient flows but in the non-canonical geometry specific to Sinkhorn flow.

7. McKean–Vlasov Diffusion Representation

There is a stochastic differential equation (SDE) whose time-marginals coincide with ρ_t:

dXt=[f/ut(Xt)g/ut(Xt(ut))+ht/ut(Xt)]dt+2(Xt/Xt(ut))dBtdX_t = [ - \partial f/\partial u_t(X_t) - \partial g/\partial u_t(X_t^{(u_t)}) + \partial h_t/\partial u_t(X_t) ]\,dt + \sqrt{2 (\partial X_t/\partial X_t^{(u_t)})}\,dB_t

where B_t is Brownian motion and (Xt/Xt(ut))=(2ut(Xt))1(\partial X_t/\partial X_t^{(u_t)}) = (\nabla^2 u_t(X_t))^{-1}. The coefficients depend on the current solution u_t of the PMA equation, making this a McKean–Vlasov (mean-field) type process. This connection is directly analogous to the relationship between the Fokker–Planck PDE and Langevin diffusion in classical potential theory.

Summary

The Sinkhorn flow emerges as the continuum limit of entropy-regularized Sinkhorn iterations in optimal transport, realized as a Wasserstein mirror gradient flow of KL divergence, with a non-canonical velocity field determined by the mirror geometry of Brenier maps. Its dynamics equivalently solve both a continuity equation with a mirror velocity and a parabolic Monge–Ampère equation for the dual potential. The flow exhibits exponential convergence under a logarithmic Sobolev inequality and can be probabilistically represented by a McKean–Vlasov diffusion. This theoretical framework clarifies the convergence of the Sinkhorn algorithm in strong topologies and establishes its PDE and stochastic process connections for a range of applications in optimal transport, generative modeling, and beyond (Deb et al., 2023).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Sinkhorn Flow.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube