Momentum-Accelerated Proximal AGD (AltGDAm)
- Momentum-Accelerated Proximal AGD (AltGDAm) is a block-coordinate method that integrates Nesterov-type and adaptive momentum for tackling nonconvex, nonsmooth composite optimization problems.
- The algorithm employs block-wise splitting, adaptive extrapolation, and monotonicity enforcement under the Kurdyka-Ćojasiewicz framework to guarantee global convergence and offer local linear rates.
- It extends classical APG techniques to applications like high-dimensional sparse regression, matrix/tensor factorization, and adversarial learning, providing provable convergence and improved empirical performance.
Momentum-Accelerated Proximal AGD (AltGDAm), also referred to as accelerated block coordinate proximal gradient with adaptive momentum, denotes a class of block-coordinate algorithms for nonconvex, nonsmooth composite optimization that incorporate Nesterov-type, adaptive, or generalized momentum into the proximal gradient or proximal linear framework. These methods combine block-wise splitting, adaptive extrapolation, and monotonicity enforcement within the structure of the Kurdyka-Ćojasiewicz (KL) convergence analysis, yielding provable global convergence and, under appropriate settings, local linear rates for a broad range of high-dimensional statistical learning and estimation problems. The approach subsumes and extends the classical accelerated proximal gradient (APG/FISTA) updates to block-separable, nonconvex, and even minimax or saddle-point settings.
1. Formal Problem Statement and Assumptions
Momentum-accelerated proximal AGD (AltGDAm) addresses composite minimization problems of the form
where the variable is partitioned into blocks, is continuous and differentiable (possibly nonconvex), and each is proper, lower semicontinuous, block-separable, and possibly nonsmooth or nonconvex (Lau et al., 2017).
The key assumptions are:
- is bounded below and admits at least one critical point , .
- Each block partial gradient is -Lipschitz in 0.
- 1 satisfies the KL property at every cluster point.
- Every block is updated at least once within any window of 2 successive steps.
This framework accommodates important regularized regression objectives (Lasso, group Lasso, capped 3, SCAD), matrix/tensor factorization with 4 penalties, robust minimax learning, and regularized image reconstruction.
2. Algorithmic Structure and Update Rules
AltGDAm iterates alternating, block-wise, momentum-accelerated proximal-gradient steps. The generic iteration includes:
Block selection (GaussâSouthwell, uniform, or cyclic rules): 5
Momentum Extrapolation (for each 6): 7
Proximal-Gradient Step (current block 8): 9
Accelerated candidate: 0
Momentum Adaptation (monotonicity enforcement): 1 The non-decreasing or shrinkage rule for the block-momentum parameter 2 prevents divergence due to momentum overshoot.
Typical step size selection is 3, 4 (Lau et al., 2017).
3. Theoretical Guarantees: Global and Local Rates
Under the above assumptions, AltGDAm exhibits the following convergence properties (Lau et al., 2017, Yang et al., 2023, Li et al., 2017):
- Global convergence: The iterates converge to a critical point 5, i.e., 6 and 7.
- Subgradient residual decay: There exists 8 such that 9, and the telescoping argument yields 0.
- KL-based local convergence rate: If 1 satisfies the KL property at 2 with exponent 3 (as for Lasso, group Lasso, SCAD), then 4 for some 5, i.e., local 6-linear convergence (Lau et al., 2017). For 7, sublinear rates 8 are implied (Yang et al., 2023, Li et al., 2017).
| Regime | KL Exponent 9 | Rate |
|---|---|---|
| Finite Steps | 0 | Finite |
| Linear | 1 | 2 |
| Sublinear | 3 | 4 |
In block-minimax contexts, momentum-accelerated Alternating GDA with proximal steps achieves improved complexity 5 for nonconvex-strongly-concave saddle point problems compared to prior 6 alternating schemes (Chen et al., 2021).
4. Connections, Variants, and Adaptive Extensions
Momentum-accelerated proximal AGD has several algorithmic variants and connections:
- Adaptive Restart and Momentum Scheduling: Algorithms such as APG-restart couple momentum with flexible restart based on function or gradient mapping tests; this prevents momentum-induced oscillations and enhances empirical convergence, maintaining global 7 rates even in the nonconvex, nonsmooth regime (Zhou et al., 2020).
- Generalized Nesterov Momentum: Extensions to 8-power momentum schemes (parametric interpolation between classical and slower-growth Nesterov momentum) provide improved robustness and high-order convergence in convex smooth preconditioned settings (Lin et al., 2024).
- Block-Coordinate Generality: AltGDAm encompasses both cyclic and random block ordering (see ABPL9 (Yang et al., 2023)) and applies to fully nonconvex, nonsmooth matrix/tensor factorization, where adaptivity and monotonicity are preserved without loss of convergence properties.
- Monotonicity Enforcement: Monotonicity checks on the objective (or surrogate) after extrapolationâaccepting only if the function value decreasesâare universal, analytically critical tools (Lau et al., 2017, Li et al., 2017).
5. Empirical and Practical Behavior
Numerical exemplars emphasize AltGDAm's advantage:
- High-dimensional sparse regression: For 0-LS, group Lasso, capped-1, and SCAD (all with 2â3 blocks), AltGDAm (with GaussâSouthwell or random block selection) consistently outperforms APG, mAPG, BPL, or APGnc in both early and late stages of convergence. The GaussâSouthwell rule, in particular, yields the fastest empirical decay of optimality gap (Lau et al., 2017).
- NMF/Tensor Decomposition: ABPL4, a practical AltGDAm implementation, demonstrates superiority over PALM, iPALM, and related block methods for 5-constrained nonnegative matrix and tensor factorization (Yang et al., 2023).
- Adversarial Learning: In robust classification minimax setups (e.g., Wasserstein robust MNIST), momentum-proximal AltGDAm accelerates both objective reduction and robust accuracy relative to GDA, prox-AltGDA, and primal-dual baselines (Chen et al., 2021).
- Preconditioned Variants: In image reconstruction, AltGDAm with EM-type preconditioning and general Nesterov momentum achieves order-optimal rates 6 in objective, robust to aggressive or conservative momentum parameters (Lin et al., 2024).
6. Extensions and Analytical Framework
- Generalization to Multiple Nondifferentiable Terms: The methodology naturally extends to settings with multiple nonsmooth separable regularizers, via block-wise fixed-point and proximal splitting (Lin et al., 2024).
- Inexact Proximal Computations: The proof template accommodates 7-approximate proximal updates, with all main results unchanged provided the error is dominated by the update distance (Yang et al., 2023, Li et al., 2017).
- KL-based Theory: All main convergence theorems fundamentally depend on the KL property, which bridges nonconvexity and non-smoothness; the method exploits sufficient decrease and subgradient bounds to invoke KL-based Lyapunov analysis, leading to precise asymptotic rates (Lau et al., 2017, Yang et al., 2023, Li et al., 2017).
7. Comparative Perspective
Relative to classic APG/FISTA frameworks, AltGDAm distinctively enables scalable, provably convergent optimization of high-dimensional nonconvex problems by:
- Supporting block-wise and adaptive update schedules.
- Allowing nonsmooth, nonconvex, and separable regularizers without bounded-domain constraints.
- Unifying monotonicity enforcement and adaptive momentum via restart/shrinkage without complex tuning requirements (Lau et al., 2017, Zhou et al., 2020, Yang et al., 2023).
- Delivering improved empirical and provable convergence over standard block proximal/alternating and non-accelerated block methods, particularly in ill-conditioned or overparameterized regimes.
The method therefore constitutes a rigorous foundation for large-scale, sparse, and structured learning in contemporary computational statistics and machine learning.