Papers
Topics
Authors
Recent
2000 character limit reached

Primal–Dual Interior-Point Framework

Updated 22 November 2025
  • Primal–dual interior-point frameworks are algorithms that solve constrained optimization problems by following a central path using both primal and dual variables.
  • They extend classical Euclidean methods to Riemannian manifolds by replacing standard derivatives with covariant derivatives, retractions, and tangent-space computations.
  • The method achieves fast local convergence—superlinear or quadratic—and robust global guarantees under standard regularity and geometric conditions.

A primal–dual interior-point framework is a class of algorithms that solve constrained optimization problems by following a trajectory (central path) through the strictly feasible region of the problem, using both primal and dual variables, via a barrier-augmented Lagrangian and perturbed Karush–Kuhn–Tucker (KKT) conditions. In the generalization to Riemannian manifolds, the framework replaces Euclidean derivatives with Riemannian objects—most notably covariant derivatives and retractions—while fully retaining the primal–dual structure, path-following, and Newton-based direction computation that underpin the success of IPMs in Euclidean domains. Central to the methodology is the use of a primal–dual Newton system constructed and updated in tangent spaces of the product manifold, maintaining interior feasibility at every step and achieving fast (superlinear, sometimes quadratic) local convergence alongside strong global guarantees under standard regularity conditions (Lai et al., 2022).

1. Riemannian Primal–Dual IPM: Mathematical Problem Statement

Let M\mathcal{M} be a dd-dimensional, connected, complete Riemannian manifold with Riemannian metric ,x\langle\cdot,\cdot\rangle_x on TxMT_x\mathcal{M}. The general nonlinear constrained optimization problem is

minxM f(x)s.t.g(x)0 Rm,h(x)=0 R,\min_{x\in\mathcal{M}}\ f(x) \qquad \text{s.t.}\quad g(x)\geq0\ \in\mathbb{R}^m,\quad h(x)=0\ \in\mathbb{R}^\ell,

with smooth (C3C^3) scalar functions ff, {gi}i=1m\{g_i\}_{i=1}^m, {hj}j=1\{h_j\}_{j=1}^\ell on M\mathcal{M}. Gradients and Hessians are taken with the Levi–Civita connection \nabla of the manifold. Slack variables sRms\in\mathbb{R}^m, so g(x)+s=0g(x)+s=0, s>0s>0, are introduced; dual multipliers z0z\geq0 for inequalities; and yy for equalities, forming the augmented variable w=(x,y,z,s)N=M×R×Rm×Rmw=(x, y, z, s)\in\mathcal{N}=\mathcal{M}\times\mathbb{R}^\ell\times\mathbb{R}^m\times\mathbb{R}^m.

The barrier-augmented Lagrangian is

Lμ(x,λ,ν)=f(x)μi=1mloggi(x)+λ,g(x)x+ν,h(x)x,\mathcal{L}_\mu(x, \lambda, \nu) = f(x) - \mu \sum_{i=1}^m \log g_i(x) + \langle \lambda, g(x)\rangle_x + \langle \nu, h(x)\rangle_x,

with barrier parameter μ>0\mu > 0.

2. Barrier-Augmented Primal–Dual Residuals and KKT System

The perturbed KKT (primal–dual) system, written as a vector field Fμ:NTNF_\mu:\mathcal{N}\rightarrow T\mathcal{N}, is

Fμ(w)=(gradxf(x)+jyjgradxhj(x)+izigradxgi(x)μigradxgi(x)gi(x) h(x) g(x)+s Zsμe)F_\mu(w) = \begin{pmatrix} \operatorname{grad}_x f(x) + \sum_j y_j \operatorname{grad}_x h_j(x) + \sum_i z_i \operatorname{grad}_x g_i(x) - \mu \sum_i \frac{\operatorname{grad}_x g_i(x)}{g_i(x)} \ h(x) \ g(x) + s \ Z s - \mu e \end{pmatrix}

where Z=diag(z),S=diag(s)Z = \operatorname{diag}(z), S = \operatorname{diag}(s), ee is the all-ones vector.

Blockwise, this generates:

  • Primal gradient residual:

rp=gradf(x)+i=1mλigradgi(x)+j=1νjgradhj(x)μi=1mgradgi(x)gi(x)=0r_p = \operatorname{grad}f(x) + \sum_{i=1}^m\lambda_i\,\operatorname{grad}g_i(x) + \sum_{j=1}^{\ell}\nu_j\,\operatorname{grad}h_j(x) - \mu\sum_{i=1}^m \frac{\operatorname{grad}g_i(x)}{g_i(x)} = 0

  • Primal equality residual: rc=h(x)=0r_c = h(x) = 0
  • Dual complementarity residual: rd=Diag(g(x))λμe=0r_d = \operatorname{Diag}(g(x))\lambda - \mu e = 0

3. Riemannian Primal–Dual Newton Step

The Newton step Δw\Delta w is computed by linearizing FμF_\mu via its covariant derivative VF(w):TwNTwNV F(w): T_w\mathcal{N}\rightarrow T_w\mathcal{N}, and solving

VF(w)[Δw]+F(w)=μeV F(w)[\Delta w] + F(w) = \mu e

in the tangent space, where ee only has nonzero component in the (s,z)(s,z) block. The Hessian block in xx is the Riemannian Hessian: HessxL(w)=Hessf(x)+j=1νjHesshj(x)+i=1mziHessgi(x)\mathrm{Hess}_x \mathcal{L}(w) = \mathrm{Hess} f(x) + \sum_{j=1}^{\ell}\nu_j\,\mathrm{Hess} h_j(x) + \sum_{i=1}^m z_i\,\mathrm{Hess} g_i(x) The full system comprises four coupled blocks; however, the condensed saddle-point system on TxM×RT_x\mathcal{M}\times\mathbb{R}^\ell is formed by block elimination: (AwHx Hx0)(Δx Δy)=(rxGS1(Zrg+μe) rh)\begin{pmatrix} A_w & H_x \ H_x^\top & 0 \end{pmatrix} \begin{pmatrix} \Delta x \ \Delta y \end{pmatrix} = \begin{pmatrix} -r_x - G S^{-1}(Z r_g + \mu e) \ - r_h \end{pmatrix} with AwA_w incorporating second derivative and scaling terms, HxH_x mapping dual search direction to the tangent space, and GG aggregating inequality constraint gradients.

The updates for Δz\Delta z and Δs\Delta s are recovered via: Δz=S1[Z(GΔx+rg)+μers],Δs=Z1[μeZΔz]\Delta z = S^{-1}[Z(G^*\Delta x + r_g) + \mu e - r_s], \quad \Delta s = Z^{-1}[\mu e - Z \Delta z]

4. Step Selection and Globalization

At each step, primal and dual variables must maintain strict positivity, enforced by a centrality condition. The step length αk\alpha_k is chosen by two rules:

  • Centrality: zk+αΔzk>0z_k + \alpha \Delta z_k > 0, sk+αΔsk>0s_k + \alpha \Delta s_k > 0
  • Armijo-type decrease on merit function φ(w)=F(w)2\varphi(w) = \|F(w)\|^2:

φ(wk+αΔwk)φ(wk)+αβgradφ(wk),Δwk\varphi(w_k + \alpha \Delta w_k) \leq \varphi(w_k) + \alpha \beta \langle \operatorname{grad} \varphi(w_k), \Delta w_k \rangle

with β(0,12)\beta \in (0, \frac{1}{2}). Backtracking reduces α\alpha until both criteria are satisfied. The update is performed via the manifold retraction: wk+1=Retrwk(αkΔwk)w_{k+1} = \operatorname{Retr}_{w_k}(\alpha_k \Delta w_k).

5. Convergence Theorems: Local and Global Guarantees

Local convergence: Given a solution (x,y,z,s)(x^*, y^*, z^*, s^*) satisfying

  • existence (A1),
  • Riemannian LICQ at xx^*,
  • strict complementarity (zi>0)(z^*_i > 0) for active ii,
  • second-order sufficiency (HessxL(w)\mathrm{Hess}_x \mathcal{L}(w^*) positive-definite on the critical subspace),

then the damped-Newton method (with diminishing μk0\mu_k \rightarrow 0 and step sizes αk1\alpha_k \rightarrow 1) converges locally superlinearly (quadratically if μk\mu_k scaled as F(wk)2\|F(w_k)\|^2 and αk1\alpha_k \rightarrow 1 rapidly) [(Lai et al., 2022), Thm 5.3].

Global convergence: Under Lipschitz continuity of FF and VFVF (under parallel transport), compact level sets, and nonsingularity of VF(w)VF(w), the line-search implementation generates iterates wkw_k for which φ(wk)0\varphi(w_k)\rightarrow 0 and every limit-point is a Riemannian KKT point [(Lai et al., 2022), Thm 6.3].

6. Algorithmic and Geometric Ingredients

  • Retraction: A mapping Retrx:TxMM\operatorname{Retr}_x: T_x\mathcal{M} \rightarrow \mathcal{M} generalizes the exponential map, satisfying Retrx(0)=x\operatorname{Retr}_x(0) = x and DRetrx(0)=ID\operatorname{Retr}_x(0) = I.
  • Vector transport: Transports tangent vectors from TxMT_x\mathcal{M} to TRetrx(η)MT_{\operatorname{Retr}_x(\eta)}\mathcal{M}, facilitating step acceptance, merit decrease, and matrix Lipschitz estimates.
  • Inner linear solves: The condensed Newton system is solved by Krylov-type methods (e.g., Conjugate Residual), exploiting operator action for efficiency and avoiding explicit dense matrix forms, essential on non-Euclidean domains.
  • Stopping criteria: Algorithm monitors F(wk)\|F(w_k)\|; termination occurs once it falls below a prescribed threshold.

7. Numerical Behavior and Applications

Empirical results show that the Riemannian primal–dual interior-point method (RIPM) achieves high accuracy and robust convergence for a variety of nonconvex optimization problems with manifold constraints (Lai et al., 2022). The generalization to the manifold setting, with correct attention to tangent-space differentiability, geometry-aware step selection, and retraction-based updates, preserves, in practice and theory, the desirable stability, fast local convergence, and global path-following guarantee of classical primal–dual IPMs. Numerical comparisons demonstrate that the method matches or exceeds performance of Euclidean competitors in problems with intrinsic manifold structure.

8. Comparison to Classical Euclidean and Other Extensions

The Riemannian framework generalizes all core steps and guarantees of the classical (Euclidean) primal–dual IPM:

  • Barrier-augmented Lagrangian and central-path system remain, now on M\mathcal{M}.
  • The Newton system incorporates Riemannian gradient and Hessian via the Levi–Civita connection.
  • Retraction and vector transport replace Euclidean vector addition and matrix products.
  • Convergence theory directly extends, modulo Riemannian KKT nonsingularity, manifold LICQ, and compactness considerations.

This formulation provides a template for further generalizations, including infinite-dimensional manifolds and optimization subject to complex geometric constraints. The robust geometric machinery enables efficient computation and makes the approach broadly applicable in modern geometric optimization (Lai et al., 2022).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Primal--Dual Interior-Point Framework.