Papers
Topics
Authors
Recent
2000 character limit reached

Dual Riemannian ADMM for Low-Rank SDPs

Updated 6 December 2025
  • The paper introduces ManiDSDP, a dual ADMM method that reformulates low-rank SDP problems using a Riemannian optimization framework on the oblique manifold.
  • It combines the global convergence of dual ADMM with efficient Riemannian trust-region subproblems, using Burer–Monteiro factorization to handle unit diagonal constraints.
  • Numerical experiments show ManiDSDP achieves high accuracy with significantly fewer iterations and improved scalability compared to traditional SDP solvers.

The dual Riemannian alternating direction method of multipliers (ADMM), as developed in "A Dual Riemannian ADMM Algorithm for Low-Rank SDPs with Unit Diagonal" (Wang et al., 4 Dec 2025), is a specialized first-order optimization algorithm for low-rank semidefinite programs (SDPs) under unit diagonal constraints. By recasting the core ADMM subproblem as Riemannian optimization over the oblique manifold via Burer–Monteiro factorization, this method—denoted ManiDSDP—combines the global convergence guarantees of dual ADMM with the empirical scalability and memory efficiency of nonlinear manifold-based search. The algorithm achieves state-of-the-art practical performance versus established SDP solvers on relaxations of both dense and sparse binary quadratic programs.

1. Problem Formulation

The target class comprises SDPs with unit diagonal and low-rank structure, typically formulated in the dual form (DSDP):

minyRmby s.t.S=A(y)C0 diag(S)=1\begin{aligned} \min_{y \in \mathbb{R}^m} \quad & b^\top y \ \text{s.t.} \quad & S = \mathcal{A}^*(y) - C \succeq 0 \ & \operatorname{diag}(S) = 1 \end{aligned}

where A:SnRm\mathcal{A} : \mathbb{S}^n \to \mathbb{R}^m is a linear map with adjoint A\mathcal{A}^*, CSnC \in \mathbb{S}^n, and bRmb \in \mathbb{R}^m are problem data. Strong duality is assumed, with the primal-dual (PSDP) and an alternative reformulation using D:=A((A)1b)D := \mathcal{A}^*( (\mathcal{A}^*)^{-1} b ):

minSSnD,S+Cs.t.S0,diag(S)=1,S=A(y)C.\min_{S \in \mathbb{S}^n} \langle D, S + C \rangle \quad \text{s.t.} \quad S \succeq 0, \quad \operatorname{diag}(S) = 1, \quad S = \mathcal{A}^*(y) - C.

This dual formulation admits tight, tractable relaxations of combinatorial problems, such as second-order relaxations for binary quadratic programs (BQPs).

2. Burer–Monteiro Factorization and the Oblique Manifold

To leverage solution low-rankness, the algorithm parameterizes SS as YYY Y^\top with YRn×pY \in \mathbb{R}^{n \times p}, pnp \ll n. The unit diagonal constraint implies YY lies on the oblique manifold:

Mp={YRn×p:Yi,:2=1  i=1,,n}\mathcal{M}_p = \{ Y \in \mathbb{R}^{n \times p} : \| Y_{i,:} \|_2 = 1 \; \forall i = 1, \ldots, n \}

The tangent space at YY is TYMp={URn×p:diag(UY)=0}T_Y \mathcal{M}_p = \{ U \in \mathbb{R}^{n \times p} : \operatorname{diag}(U Y^\top) = 0 \}, with induced Riemannian metric U,VY=tr(UV)\langle U, V \rangle_Y = \operatorname{tr}(U^\top V). The orthogonal projection is PY(U)=U(UY)YP_Y(U) = U - (U Y^\top) Y.

This factorization converts positive semidefiniteness and unit-diagonal constraints into a nonlinear manifold constraint, for which Riemannian optimization techniques are natural.

3. Augmented Lagrangian and Dual ADMM Decomposition

The ManiDSDP approach constructs an augmented Lagrangian in primal-dual variables (S,y,X~)(S, y, \widetilde X):

Lσ(S,y,X~)=D,S+CX~,A(y)SC+σ2A(y)SCF2\mathcal{L}_\sigma(S, y, \widetilde X) = \langle D, S + C \rangle - \langle \widetilde X, \mathcal{A}^*(y) - S - C \rangle + \frac{\sigma}{2} \| \mathcal{A}^*(y) - S - C \|_F^2

ADMM alternates:

  • S-update: Sk+1=argminS0,diag(S)=1Φk(S)S^{k+1} = \arg \min_{S \succeq 0, \operatorname{diag}(S) = 1} \Phi_k(S), where Φk(S)=Lσk(S,yk,X~k)\Phi_k(S) = \mathcal{L}_{\sigma_k}(S, y^k, \widetilde X^k).
  • y-update (closed form): yk+1=(A)1(Sk+1+C)y^{k+1} = (\mathcal{A}^*)^{-1}(S^{k+1} + C).
  • Multiplier update: X~k+1=X~kσk(A(yk+1)Sk+1C)\widetilde X^{k+1} = \widetilde X^k - \sigma_k ( \mathcal{A}^*(y^{k+1}) - S^{k+1} - C ).

Dual variables are recovered as zk+1=diag((X~k+1+D)Sk+1)z^{k+1} = \operatorname{diag}( (\widetilde X^{k+1} + D)\, S^{k+1}), Xk+1=X~k+1+Ddiag((X~k+1+D)Sk+1)X^{k+1} = \widetilde X^{k+1} + D - \operatorname{diag}( (\widetilde X^{k+1} + D) S^{k+1} ).

4. Riemannian Subproblem and Trust-Region Solver

The S-update is restated over Mp\mathcal{M}_p via S=YYS = Y Y^\top, yielding

minYMpΨk(Y):=Φk(YY)\min_{Y \in \mathcal{M}_p} \Psi_k(Y) := \Phi_k(YY^\top)

This subproblem is solved with a Riemannian trust-region method. Let G=Φk(S)=X~kσk(A(yk)SC)+DG = \nabla \Phi_k(S) = \widetilde X^k - \sigma_k ( \mathcal{A}^*(y^k) - S - C ) + D, and X=Gdiag(GS)X = G - \operatorname{diag}(GS). The Riemannian gradient is $2 X Y$ and the projection formula above applies.

Second-order optimality checks are performed during the subsolve: the trust-region iteration stops when XYϵk\|X Y\| \leq \epsilon_k, λmin(X)τk\lambda_{\min}(X) \geq -\tau_k, with ϵk,τk0\epsilon_k, \tau_k \to 0, ϵk<\sum \epsilon_k < \infty, τk<\sum \tau_k < \infty. On encountering negative curvature, eigendirections VV of XX are used to augment YY and escape saddles.

The overall ManiDSDP (outer ADMM) algorithm includes adaptation of rank pp, penalty σ\sigma, and an optional saddle-escape step.

5. Theoretical Guarantees

Convergence analysis (Theorem 6.1 of (Wang et al., 4 Dec 2025)) establishes that, under the inexactness schedules ϵk<\sum \epsilon_k < \infty, τk<\sum \tau_k < \infty and bounded multipliers, every cluster point of ManiDSDP iterates satisfies the KKT conditions of the dual SDP. The core analytical lemmas link descent of the augmented Lagrangian and reduction of the feasibility residual Rk2\|R^k\|^2, leading to Rk2<\sum \|R^k\|^2 < \infty and Rk0R^k \to 0, with asymptotic positivity and slackness properties for XkX^k.

A plausible implication is that, provided the Riemannian subproblems are solved to sufficiently high accuracy, the algorithm approaches optimality even in the presence of the nonconvex factorization.

6. Computational Cost and Scalability

The dominant per-iteration costs are as follows:

  • Forming G=Φk(S)G = \nabla \Phi_k(S): O(nnz(A)+n2)O(\text{nnz}(\mathcal{A}) + n^2).
  • Riemannian trust-region solve: per step, cost is O(np2+cost(A),A)O(n p^2 + \text{cost}(\mathcal{A}), \mathcal{A}^*) for gradient and Hessian-vector products. Since pnp \ll n, total cost per outer iteration is O(np2+nnz(A))O(n p^2 + \text{nnz}(\mathcal{A})).

Typically, the number of trust-region steps per ADMM iteration is in the tens. This cost structure is highly favorable for large, low-rank problems.

7. Numerical Results and Algorithmic Phenomena

Comprehensive experiments benchmark ManiDSDP against MOSEK, COPT, SDPNAL+, and ManiSDP on second-order SDP relaxations of dense and sparse BQPs. For dense BQPs of size up to q=80q=80, ManiDSDP completes in 10210^210310^3 s, outperforming MOSEK (out-of-memory), COPT (numerical issues), SDPNAL+ (time 103\sim 10^3 s), and ManiSDP (twice larger time). On sparse BQPs ((q,t)=(20,120)(q,t)=(20,120)), ManiDSDP completes in 10210^2 s, while other solvers often time out.

ManiDSDP achieves 101210^{-12}101510^{-15} accuracy in a few tens of outer ADMM iterations, compared to 104+10^4+ iterations for traditional ADMM. Maximal factorization sizes and outer iteration counts are lower than for ManiSDP.

A notable “residue-diving” phenomenon is observed: the KKT residual log10ηmax_{10} \eta_{\max} can drop abruptly from 12-1 \sim -2 to 1215-12 \sim -15 in a single ADMM step. This suggests a rapid improvement in optimality and feasibility at late stages.

The overall method is shown to be both memory- and time-efficient, scalable, and competitive for challenging SDPs with unit diagonal structure (Wang et al., 4 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Dual Riemannian Alternating Direction Method of Multipliers (ADMM).