Progressive Power Homotopy (Prog-PowerHP)
- Progressive Power Homotopy (Prog-PowerHP) is a homotopy-based optimization paradigm that sequentially deforms complex non-convex problems into tractable surrogates using power transformations and Gaussian smoothing.
- It employs adaptive scheduling in both zeroth- and first-order regimes to ensure iterates remain in the basin of attraction, enhancing robustness against ill-conditioning and multi-modality.
- Empirical results show state-of-the-art performance in tasks like adversarial attacks, phase retrieval, and constrained optimization in power systems.
Progressive Power Homotopy (Prog-PowerHP) is a homotopy-based optimization paradigm for navigating high-dimensional non-convex objectives. The central principle is to deform a target (potentially non-convex and ill-conditioned) problem into a sequence of intermediate, more tractable surrogates through the simultaneous application of power transformations and smoothing, or related structured relaxations. As the homotopy path is traversed, algorithmic iterates are tracked in the parameter space, ensuring they remain in the basin of attraction of progressively less regularized, more challenging surrogates, culminating close to the true global optimum. Prog-PowerHP has been developed in both zeroth- and first-order settings, with substantive theoretical justification and empirical validation in machine learning and scientific optimization tasks (Xu, 17 Nov 2025, Xu, 22 Jan 2026).
1. Power-Transformed Gaussian-Smoothing: Surrogate Construction
For unconstrained maximization , Prog-PowerHP introduces the power-transformed, Gaussian-smoothed surrogate: where is the power parameter and is the smoothing scale (Xu, 17 Nov 2025). The power transform exponentially "tilts" the distribution toward regions of high objective value, causing the maxima of to cluster near that of as . The Gaussian smoothing regularizes the landscape, suppressing spurious local maxima and facilitating robust gradient estimation.
A key insight is the contrast to directly smoothing . Smoothing , even with modest , yields surrogate maxima that are globally attracted to (the global maximizer of ), rather than merely local optima. This mechanism is robust to ill-conditioning and multi-modality in , provided and are appropriately scheduled.
2. Homotopy/Continuation and Scheduling Strategies
Prog-PowerHP employs a single-loop homotopy scheme in which either the smoothing parameter () is geometrically decayed toward a fixed lower bound (zeroth-order regime (Xu, 17 Nov 2025)), or both the power parameter () and smoothing scale () are progressively adjusted (first-order regime (Xu, 22 Jan 2026)):
- Zeroth-Order: fixed; , for decay rate and floor .
- First-Order: ; for schedule with .
This scheduling forms the Prog-PowerHP "homotopy path"—iterates are initialized in the tractable regime (large , small ), then tracked as the surrogate sharpens, ensuring solutions at each stage are in the basin of attraction for the next surrogate.
In constrained or structured settings such as AC-OPF, the homotopy includes both objective and constraint relaxations parameterized by a schedule , interpolating from convex, relaxed subproblems to the original non-convex instance (Li et al., 11 Nov 2025). This allows policy learning procedures to traverse from trivial to challenging regimes under self-supervision.
3. Gradient Estimation and Algorithmic Framework
Zeroth-Order Optimization
The method employs the Gaussian-smoothing gradient estimator: with (Xu, 17 Nov 2025). This estimator satisfies , so dividing by yields an unbiased estimate. Variance is controlled and bounded (Lemma 4.2 in (Xu, 17 Nov 2025)).
First-Order Stochastic Optimization
For models with parametric structure and explicit gradients: where , (Xu, 22 Jan 2026). Iterates are updated via with carefully chosen step-sizes ().
Constrained Policy Learning (AC-OPF)
The policy is trained to minimize a homotopically evolving penalty objective: where denote (relaxed) equality/inequality constraints, and penalties , (Li et al., 11 Nov 2025).
4. Theoretical Properties and Global Behavior
Theoretical analysis establishes strong convergence behavior under mild regularity assumptions:
- For appropriate homotopy schedule and sufficiently large , stationary points of can be made to lie arbitrarily close to the global maximizer (Xu, 17 Nov 2025, Xu, 22 Jan 2026).
- Under stepsizes , the iteration complexity to reach is (for ).
- In constrained settings, empirical evidence demonstrates strong feasibility restoration and generalization to unseen instances, though the NP-hardness of global optimization in the general AC-OPF setting precludes formal global guarantees (Li et al., 11 Nov 2025).
Phase transition analyses in tensor PCA show a bifurcation in landscape geometry along the homotopy path: at sufficiently large smoothing, the problem becomes strictly convex with a unique global maximizer, while as smoothing decreases, secondary local maxima emerge, justifying the necessity for a carefully managed homotopy schedule (Anandkumar et al., 2016).
5. Empirical Performance and Benchmark Results
Prog-PowerHP has demonstrated:
- Robust top-three ranking on standard non-convex test functions (e.g., Ackley, Rosenbrock) among a pool of eight zeroth-order and homotopy algorithms (Xu, 17 Nov 2025).
- State-of-the-art results in adversarial black-box attacks (ImageNet, ), outperforming GS-PowerOpt, ZO-SLGH, CMA-ES, Square Attack, etc., in terms of both success rate and perturbation size.
- Significant improvement in phase retrieval at sample-to-dimension ratios approaching the information-theoretic limit, with success rates as high as 92% () compared to 41–56% for strong first-order baselines (Xu, 22 Jan 2026).
- In training of under-parameterized two-layer ReLU networks, lower test error and higher success rates than SGD, Adam, and various specialized landscape-smoothing methods (Xu, 22 Jan 2026).
- For parametric AC-OPF, sustained feasibility above 98% on unseen load settings (75–150% of nominal), and improved objective values relative to penalty-based baselines (Li et al., 11 Nov 2025).
| Scenario | Prog-PowerHP Result | Reference |
|---|---|---|
| ImageNet Attack | 1st place; best trade-off (d=150,528) | (Xu, 17 Nov 2025) |
| Phase Retrieval | 0.07 rel. error (92% succ., d=100) | (Xu, 22 Jan 2026) |
| AC-OPF (30-bus) | 1.7e–3 mean eq, 0.00 mean ineq violation | (Li et al., 11 Nov 2025) |
6. Variants, Extensions, and Limitations
While originally formulated in a zeroth-order framework for black-box optimization, Prog-PowerHP admits generalization to first-order and policy-learning contexts. Notable aspects include:
- The power-homotopy principle (progressive exponentiation of the objective) and Gaussian smoothing (mollification) are jointly responsible for global-attracting surrogates.
- Adaptive scheduling (progressively increasing or decaying ) enhances numerical stability and empirical recovery rates.
- Puiseux series and tropical-algebraic preprocessing allow extension to singularities and multiple root tracking in algebraic homotopy (Bliss et al., 2016).
The principal limitations are the need for hyperparameter tuning (), sampling cost scaling as per iteration, and, in some settings, lack of formal global guarantees due to intractability of the underlying problem (notably in constrained or high-dimensional combinatorial regimes).
7. Connections, Impact, and Outlook
Prog-PowerHP unifies a variety of homotopy- and smoothing-based approaches:
- In tensor PCA, it provides a rigorous phase transition analysis and nearly-linear time global recovery for spiked tensors at the optimal statistical threshold (Anandkumar et al., 2016).
- In policy learning for control and power systems, it achieves scalable constraint-aware optimization without labeled solutions via structured objective and constraint homotopies (Li et al., 11 Nov 2025).
- The method matches or surpasses the empirical performance of state-of-the-art global optimization heuristics, with superior robustness in high dimensions and cluttered landscapes (Xu, 17 Nov 2025, Xu, 22 Jan 2026).
A plausible implication is that Prog-PowerHP, when appropriately tailored, provides a principled and flexible toolkit for diverse non-convex search problems, and sets a template for future research integrating power-based homotopies, smooth surrogate design, and model-aware continuation schemes.