Restarted Halpern PDHG Method
- The paper introduces rHPDHG, a matrix-free, primal-dual method that achieves accelerated two-stage linear convergence for large-scale linear programming and saddle-point problems.
- The method leverages Halpern fixed-point iterations with an outer restarting schedule and reflection steps to reduce dependence on pessimistic global condition constants.
- GPU-aware implementations and adaptive parameter tuning enhance its practical performance, ensuring robust convergence in both feasible and infeasible LP settings.
The Restarted Halpern Primal-Dual Hybrid Gradient (rHPDHG) Method is a matrix-free, primal–dual first-order optimization algorithm designed to solve large-scale linear programming (LP) and related saddle-point problems. By incorporating Halpern-type fixed-point iterations, an outer restarting schedule, and—in some variants—a reflection step, rHPDHG achieves accelerated two-stage linear convergence, improved practical performance, and efficient parallelization. This article delineates the algorithmic principles, theoretical foundations, condition measures, extensions, implementation details, and empirical performance of rHPDHG and related advances.
1. Algorithmic Foundation and Structure
The rHPDHG method addresses LPs of the form: using a primal–dual splitting approach. The core innovation over classical PDHG algorithms is the adoption of Halpern fixed-point anchoring. Specifically, within each inner epoch (indexed by ), the update is: where denotes the standard PDHG operator and %%%%2%%%% is the anchor point (typically the initial point or the output of a prior epoch). After a prescribed number of inner iterations, a restart is triggered by re-anchoring to the current best iterate. This process is repeated across outer epochs, yielding a doubly-nested iterative scheme.
A further enhancement, termed reflected rHPDHG (or ), applies Halpern iteration to the reflection of the PDHG operator, i.e., using in place of , thereby taking more aggressive steps.
The restart criterion can be fixed frequency or adaptive, based on monitoring the fixed-point residual: where is a parameter matrix scaling the norm. Adaptive rules include restarting upon sufficient decrease of the residual, lack of local progress, or reaching a maximum inner-epoch length.
2. Theoretical Convergence and Two-Stage Behavior
A rigorous analysis of rHPDHG establishes global accelerated linear convergence for feasible and bounded LPs. In particular, for a restart period (with a sharpness parameter), the following holds for the restart iterates: leading to an iteration complexity for target accuracy (Lu et al., 23 Jul 2024).
The convergence admits a refined two-stage characterization:
- Stage I: Active Set Identification—The algorithm identifies the correct support structure by partitioning variable indices into nonbasic (), nondegenerate basic (), and degenerate (). The duration depends on the nondegeneracy metric
resulting in a complexity , with a local sharpness constant (Lu et al., 23 Jul 2024).
- Stage II: Local Linear Convergence—Once identification is achieved, the algorithm enjoys accelerated local linear convergence with a rate dictated by another local sharpness , and complexity .
This refined analysis replaces prior global Hoffman-type constants—often exponentially pessimistic—with practical, locally sharp rates (Lu et al., 2023).
3. Condition Measures: Sharpness, Limiting Error Ratio, and Geometric Quantities
The complexity of rHPDHG is fundamentally governed by intrinsic condition measures:
- LP sharpness ( and ) quantifies the sensitivity of the objective to deviation from the optimum, capturing how rapidly the objective increases away from the optimal set (Xiong et al., 2023, Xiong et al., 2023).
- Limiting Error Ratio measures local geometric feasibility properties, defined as
where is the feasible set. The limiting value as informs algorithm complexity near the optimum.
- Geometric Condition Numbers such as in (Xiong, 5 Oct 2024), defined explicitly from the optimal basis and solution, allow for accessible, closed-form iteration bounds.
A central insight is the reciprocal relationship between sharpness and the iteration bound: as the LP becomes "sharper" (larger ), the algorithm converges more rapidly. The two-stage complexity splits accordingly: an initial phase determined by sharpness and stability under data perturbation, followed by a faster, locally conditioned phase controlled by the minimal nonzero solution component and the conditioning of the optimal basis matrix.
4. Handling Infeasibility and Certificates
rHPDHG, like its base PDHG, is effective for both feasible and infeasible LPs. For infeasible cases, the method converges to the infimal displacement vector (the minimum-norm element of ). The updated Halpern and reflection strategies yield accelerated linear convergence in extracting infeasibility certificates, even in the absence of strong regularity conditions (Lu et al., 23 Jul 2024, Jiang et al., 2023). This accelerates and stabilizes the identification of invalid LP instances.
5. Extensions: Reflective and Parameter-Free Variants
The reflected rHPDHG extension () applies Halpern averaging with the reflected operator , reducing the required inner-epoch length by half and improving per-epoch contraction rates: All convergence guarantees of rHPDHG extend, with constants improved by a factor of 2 (Lu et al., 23 Jul 2024, Lu et al., 18 Jul 2025).
Related methods, such as adaptive PDHG with automatic stepsize tuning (Goldstein et al., 2013) and relaxed, parameter-free variants with line search (McManus et al., 21 Mar 2025), provide alternate avenues for removing user input dependence on problem constants, using residual balancing or operator-based line searches.
6. GPU-Aware Implementation and Empirical Evaluation
The matrix-free, matrix–vector-multiplication structure of rHPDHG makes it ideal for GPU acceleration. Reference implementations (e.g., HPDLP in Julia/CUDA (Lu et al., 23 Jul 2024), cuPDLP+ (Lu et al., 18 Jul 2025)) exploit parallelism and tailored restart/anchor management. In large-scale tests on MIPLIB LP relaxations:
- rHPDHG and realize per-instance speedups of – over previous PDHG-based solvers at moderate accuracy, and – at high accuracy or with presolve (Lu et al., 18 Jul 2025).
- The PID-controlled primal weight update in cuPDLP+ dynamically balances primal and dual progress by updating the weight using a logarithmic-scale PID controller:
where is the log-scale ratio of primal and dual residuals (Lu et al., 18 Jul 2025).
- Constant stepsize rules (e.g., ) remove the need for per-iteration spectral estimation.
- Empirical performance matches or exceeds that of existing GPU-based first-order solvers; e.g., cuPDLP+, incorporating r2HPDHG, achieves substantial gains in solve time and robustness.
7. Practical Guidelines and Perspective
For effective application of rHPDHG:
- Exploit row preconditioning and tuned stepsize ratios to moderate the iteration bound’s dependence on condition numbers and sharpness (Xiong et al., 2023).
- Use adaptive or residual-based restart criteria to maintain fast local convergence after identification of the optimal structure (Lu et al., 23 Jul 2024, Lu et al., 18 Jul 2025).
- Apply reflection (r2HPDHG) to further speed up contraction, especially in ill-conditioned regimes.
- Warm-start or combine with learning-based approaches (e.g., PDHG-Net) for very large instances or in hybrid architectures (Li et al., 4 Jun 2024).
- For challenging test cases, manipulate the disparity ratio of the optimal variables to control the iteration count in Stage I (Xiong, 1 Jan 2025).
In summary, the restarted Halpern PDHG method unifies fixed-point averaged iteration, restart-based acceleration, local geometric conditioning, and scalable, parallel-first-order optimization into a framework with theoretically optimal and empirically robust convergence for large-scale LP and convex saddle-point problems. Recent advances in solution certification, parameter-free variants, step-size scheduling, and GPU-tailored implementations have made rHPDHG and its extensions competitive across diverse domains and datasets.