Discrete Optimal Transport
- Discrete Optimal Transport studies the optimal reallocation of discrete masses under cost minimization with prescribed marginal constraints.
- It employs variational and polyhedral formulations, using convex functions and power diagrams to derive robust, scalable numerical algorithms.
- DOT is applied in computational geometry, statistical imaging, and machine learning, offering efficient methods for large-scale transport problems.
Discrete optimal transport (DOT) is the paper of optimal transportation problems where measures are discrete, i.e., supported on finite sets or domains partitioned into finitely many regions. In DOT, the goal is to minimize a prescribed transportation cost when reallocating "mass" from one distribution to another, subject to prescribed marginal constraints. DOT stands at the interface of convex geometry, combinatorial optimization, computational geometry, and modern applications in statistics, data science, and computer vision. The domain supports distinct variational principles, linear programming formulations, deep connections to power diagrams and convex polytopes, and leads to scalable algorithms for applications involving large discrete domains.
1. Variational and Polyhedral Formulations of Discrete Optimal Transport
DOT is grounded in a variational framework that finds its most geometric statement in the context of convex subdivision and power diagrams. Given a compact convex domain , a positive density on , a set of target points in , and prescribed masses with , one seeks a partition of into convex regions such that for all (Gu et al., 2013).
The partition is encoded via a "height vector" :
and via the convex, piecewise linear function:
The energy functional
is strictly convex (on the hyperplane ), -smooth, and admits an explicit Hessian. Its gradient is the vector of cell -masses: . Thus, finding such that the prescribed masses are recovered is equivalent to solving . The unique solution (modulo additive constant) can be efficiently computed (e.g., via Newton’s method) (Gu et al., 2013).
From the duality perspective, in pure discrete settings, optimal transport is classically cast as a linear program: for empirical measures on finite sets , ,
where is a prescribed ground cost and the probability vectors and specify the mass at and respectively (Solomon, 2018).
2. Geometric Structures: Power Diagrams and Links to Convex Geometry
A major structural insight is that the regions —which partition for the DOT problem in the semidiscrete or continuous-to-discrete setting—are exactly the cells of a power diagram (weighted Voronoi diagram) defined by points with . The connection is precise:
Thus, DOT can be reframed as adjusting the power diagram weights so that induced cell volumes match prescribed masses (Gu et al., 2013). This geometric viewpoint underpins both the variational approach and algorithmic strategies, leading to direct links to discrete convex polytopes (e.g., Minkowski-type problems) and to the discrete Monge–Ampère equation via the volume of the convex hull of neighboring gradients.
3. Algorithmic Approaches: Convex Optimization and Multi-Scale Methods
The variational framework makes the DOT problem amenable to efficient numerical algorithms. The strictly convex function (with explicit gradient and Hessian) yields, upon restricting to , a unique solution which is accessible by Newton’s method, with explicit computation of gradients and Hessians on each iteration. The Hessian for is:
where is the shared codimension-1 face between adjacent cells.
For fully discrete problems, linear programming approaches remain standard, but recent advances exploit sparsity and geometric structure. Hierarchical multi-scale algorithms, such as the sparse shielding neighborhood method, construct local approximations and coarse-to-fine initializations: only a sparse subset is handled at each iteration, leading to dramatic reductions in run-time and memory usage scaling as . These methods verify optimality by leveraging geometric shielding conditions on the dual constraints and are the foundation of performant solvers for large DOT problems (Schmitzer, 2015, Rauch et al., 28 Feb 2025).
In applied contexts, entropic regularization (Sinkhorn algorithm) provides scalable, highly parallelizable DOT solvers, though they introduce regularization error and may fail to achieve high precision in practice (Solomon, 2018, Lu et al., 29 Jul 2024).
4. Fundamental Connections: Discrete Monge–Ampère Equation and Convex Duality
The discrete Monge–Ampère equation (DMAE) emerges naturally from the variational DOT framework. Whereas the classical Monge–Ampère equation constrains the volume element of the Hessian determinant of a convex function, the discrete analogue defines a “discrete Hessian determinant” at each target as the volume of the convex hull formed by neighboring gradients. The key mass-balance condition,
is equivalent to requiring the discrete Hessian determinant at equals . This fact directly relates DOT to combinatorial convex geometry and supports novel variational proofs of classical rigidity theorems, including Alexandrov's theorem on the realization of polytopal metrics (Gu et al., 2013).
Moreover, the DOT variational principle offers a unified convex optimization view, encompassing Minkowski-type problems for polytopes and providing new, structurally explicit proofs for results in discrete differential geometry.
5. Applications and Algorithmic Implications
DOT and its geometric/variational formulation have several practical implications:
- Computational geometry: Efficient algorithms for power diagrams enable fast solutions to the DOT problem in high dimensions, especially in applications involving spatial partitioning, such as mesh generation, centroidal Voronoi tessellations, and tessellation-based quantization (Gu et al., 2013).
- Numerical optimal transport: The ability to compute explicit gradients and Hessians in the -weights makes the Newton-based approach highly efficient, particularly when compared to generic, dense linear programming solvers for large .
- Scalable computation: Multi-scale and shielding techniques enable DOT computations for massive datasets (e.g., in statistical imaging, machine learning, and computational graphics) (Schmitzer, 2015, Rauch et al., 28 Feb 2025).
- Statistical and data science applications: DOT underwrites the computation of Wasserstein barycenters, domain adaptation, and distributional clustering in the fully discrete or discretized setting (Anderes et al., 2015).
6. Generalizations and Extensions
The discrete framework for optimal transport not only solves specific allocation problems but also generalizes to:
- Wasserstein barycenters in the discrete setting, where support can be restricted to candidate centroids and sparsity and non-mass-splitting transport properties hold (Anderes et al., 2015).
- Discrete Monge–Ampère-type equations, with rigidity and convergence properties accessible through the variational framework.
- Combinatorial analogues on graphs and higher-dimensional polytopes, incorporating the interplay of linear programming duality, convex geometry, and discrete analysis.
- Extensions to settings with one measure supported on continuous curves or polylines (“3/4‑discrete” OT) and to hybrid discrete–continuous scenarios (Gournay et al., 2018).
7. Theoretical and Algorithmic Impact
The variational principles for DOT and the identification of discrete optimal maps with power diagram adjustments provide a unifying structure. They yield direct geometric and algorithmic interpretations, robust guarantees on uniqueness and stability, and enable the derivation and analysis of efficient, large-scale solvers. The DOT variational approach delivers not just numerical efficiency, but also theoretical advancements—such as new proofs for old theorems, and a rigorous bridge between discrete geometry, classical analysis, and combinatorial optimization.
This synthesis of convex variational principles, geometric structure, and computational tractability defines discrete optimal transport as a foundational methodology in modern geometric analysis, computational mathematics, and data science.