Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 102 tok/s
Gemini 2.5 Pro 58 tok/s Pro
GPT-5 Medium 25 tok/s
GPT-5 High 35 tok/s Pro
GPT-4o 99 tok/s
GPT OSS 120B 472 tok/s Pro
Kimi K2 196 tok/s Pro
2000 character limit reached

Convex Conic Optimization

Updated 3 September 2025
  • Convex conic optimization is a framework that models convex problems using closed cone constraints, covering linear, quadratic, semidefinite, and exponential programming.
  • It leverages unified primal–dual optimality theory and employs algorithms such as interior-point, augmented Lagrangian, and operator splitting methods for efficient solution.
  • Its diverse applications span control, signal processing, finance, and machine learning, with techniques adapted for large-scale, structured problems.

Convex conic optimization encompasses the paper and solution of convex optimization problems where some or all constraints require a vector or matrix variable to lie in a closed convex cone. These problems are simultaneously characterized by the interplay between algebraic structure and geometric constraints, and they form a rich subclass capturing linear, quadratic, semidefinite, and exponential programming as special cases. The class admits unifying primal–dual optimality theory, and supports both interior-point and first-order algorithms that are central to numerical optimization, control, signal processing, operations research, and emerging fields including machine learning and robust optimization.

1. Problem Class and Mathematical Formulation

A convex conic optimization problem is typically formulated as

minx p(x)=ρ(x)+γ(x)s.t.AxbK, xχ\min_{x} \ p(x) = \rho(x) + \gamma(x) \quad \text{s.t.} \quad Ax - b \in \mathcal{K}, \ x \in \chi

where ρ\rho is a closed convex (possibly nonsmooth) function, γ\gamma is convex with a Lipschitz continuous gradient, AA is a real matrix, bb a given vector, K\mathcal{K} a closed convex cone in a finite-dimensional space, and χ\chi a “simple” convex compact set. The feasible set is the intersection of a convex cone and a compact convex subset, as in

{xχ:AxbK}.\{ x \in \chi: Ax - b \in \mathcal{K} \}.

Conic constraints encapsulate a broad array of standard problems:

  • In linear programming, K=R+m\mathcal{K} = \mathbb{R}_+^m.
  • In second-order cone programming (SOCP), K\mathcal{K} is a Cartesian product of Lorentz cones.
  • In semidefinite programming (SDP), K\mathcal{K} consists of positive semidefinite matrices.
  • Additional cones include the exponential and power cones, which model entropy and geometric mean type constraints.

A penalty Lagrangian formulation arises naturally: Lμ(x,y)=p(x)+μ2dK(Axby/μ)212μy2,\mathcal{L}_{\mu}(x, y) = p(x) + \frac{\mu}{2} d_{\mathcal{K}}(Ax-b-y/\mu)^2 - \frac{1}{2\mu} \|y\|^2, where yy is a dual multiplier and dK()d_{\mathcal{K}}(\cdot) denotes Euclidean distance to K\mathcal{K} (Aybat et al., 2013). This framework supports both Lagrangian and augmented Lagrangian approaches.

2. Core Theoretical Framework

The theoretical properties of convex conic optimization problems mirror and extend classic convex optimization duality. Strong duality and KKT conditions hold under standard regularity (Slater-type) conditions, with dual cones K\mathcal{K}^* defining the dual problem. The dual function in augmented Lagrangian methods incorporates the Moreau envelope of the indicator of K\mathcal{K}, regularizing non-smooth penalty terms (Aybat et al., 2013).

Conic duality plays a foundational role not only for optimality characterization, but also for the design of numerical methods (both first-order and interior-point). Splitting algorithms such as Douglas–Rachford and Spingarn’s method exploit primal–dual structure in a Hilbert space via variable replication and projections (e.g., in decomposing sparse SDPs through clique-based chordal sparsity) (Sun et al., 2013).

3. Algorithmic Paradigms

Algorithmic strategies for convex conic optimization span interior-point, operator splitting, and augmented Lagrangian frameworks.

  • First-Order Augmented Lagrangian Methods: The ALCC algorithm iteratively solves an inexact penalized subproblem over the feasible (compact) set, followed by a dual update via projection onto the cone. Subproblems are tackled by accelerated proximal gradient (APG) methods (e.g., FISTA variants). The complexity is O(log(1/ϵ))O(\log(1/\epsilon)) outer iterations and O(ϵ1log(1/ϵ))O(\epsilon^{-1} \log(1/\epsilon)) total oracle calls to reach ϵ\epsilon-optimality and ϵ\epsilon-feasibility (Aybat et al., 2013).
  • Splitting Algorithms and Operator-Based Methods: ADMM techniques are applied to homogeneous self-dual embeddings to solve feasibility forms of the conic KKT system, yielding algorithms like SCS that are parallelizable and amenable to large-scale instances (O'Donoghue et al., 2013); similar strategies underpin SuperSCS and ADMM-based COSMO (Sopasakis et al., 2019, Garstka et al., 2019).
  • Interior-Point Methods with Barrier Functions: Classical primal–dual path-following methods rely on self-concordant and logarithmically homogeneous barrier functions on the cone. The iteration complexity is O(νlog(1/ϵ))O(\sqrt{\nu}\log(1/\epsilon)), with ν\nu the barrier parameter (Papp et al., 2021, Papp et al., 22 Feb 2025). For nonsymmetric cones (e.g., hyperbolicity cones, power and exponential cones), modern implementations such as alfonso adapt interior-point techniques using user-supplied oracles for the barrier gradient and Hessian, not requiring symmetric cone structure (Papp et al., 2021, Papp et al., 22 Feb 2025).
  • Decomposition and Chordal Techniques: For structured SDPs and conic problems, exploitation of partial separability via chordal decomposition accelerates projection and factorization steps, as in COSMO and Clarabel (Sun et al., 2013, Goulart et al., 21 May 2024).
  • Composite and Multi-block ADMM: For problems with block-separable structure, generalized ADMM with semi-proximal regularization and overrelaxation can efficiently handle DNN-SDPs and related large-scale formats (Xiao et al., 2015).
  • First-Order Gradient-type Primal–Dual Methods: Recent work on proportional–integral projected gradient methods (PIPG) and their extrapolated variants (xPIPG) achieves improved rates for both primal–dual gap and constraint violation, with proven infeasibility detection and acceleration through extrapolation (Yu et al., 2021, Yu et al., 2022).

4. Complexity, Accuracy, and Scalability

Theoretical guarantees for convex conic methods are rooted in the interplay between the structure of the cone, the smoothness of the objective, and the conditioning of the constraints:

  • Augmented Lagrangian (ALCC) achieves O(log(1/ϵ))O(\log(1/\epsilon)) outer iterations for ϵ\epsilon-accuracy, with each iteration cost proportional to the cost of a proximal mapping or projection (Aybat et al., 2013).
  • First-order splitting and ADMM require fewer per-iteration computations, exploit data sparsity, and can process problems with up to a billion nonzeros (O'Donoghue et al., 2013), but may suffer from slow tail-convergence.
  • Interior-point methods (both classic and modern, including those for nonsymmetric cones) achieve optimal iteration complexity with robust convergence and high accuracy, but have higher per-iteration computational cost due to matrix factorizations; scalability is improved through chordal SDP decomposition (Papp et al., 2021, Goulart et al., 21 May 2024).

Modern methods provide automatic infeasibility detection (through homogeneous embedding or dynamic residual checks) and allow for certificates of optimality or unboundedness alongside solution recovery (O'Donoghue et al., 2013, Yu et al., 2022).

5. Applications and Model Structures

Convex conic optimization serves as the computational core for a diverse range of applied and theoretical contexts:

  • Semidefinite Programs: Robust control, matrix completion, combinatorial relaxations, and quantum information are expressed via SDPs, leveraging product cone and chordal representations (Aybat et al., 2013, Sopasakis et al., 2019, Goulart et al., 21 May 2024).
  • Compressed Sensing and Signal Processing: Structure-exploiting algorithms, especially those based on first-order methods, address large-scale problems where shrinkage and projection are computational bottlenecks (Aybat et al., 2013, O'Donoghue et al., 2013).
  • Portfolio Optimization and Finance: Second-order cones model risk, and large-scale conic solvers efficiently handle long-only and mean-variance models (O'Donoghue et al., 2013, Goulart et al., 21 May 2024).
  • Optimal Control and Real-Time MPC: First-order conic solvers underpin predictive control for systems with high-dimensional state and input constraints, especially via methods designed for embedded and GPU acceleration (Yu et al., 2021, Yu et al., 2022, Kamath et al., 2022, Chen et al., 26 Dec 2024).
  • Machine Learning and Quantum Information: CGP (Conic Geometric Programming) generalizes GPs and SDPs to accommodate exponential and entropy-type constraints, with direct impact in approximations of the permanent, hitting-time estimation, and quantum channel capacity (Chandrasekaran et al., 2013).

Modeling frameworks such as Convex.jl and downstream solvers like Clarabel and COSMO allow users to specify convex conic problems in high-level, DCP-compliant languages, automatically parse structure, and select appropriate solution routines (Udell et al., 2014, Goulart et al., 21 May 2024, Garstka et al., 2019).

6. Algorithmic Comparison and Implementation Considerations

Comparisons among algorithms sharpen understanding of trade-offs:

  • Augmented Lagrangian methods (ALCC) yield competitive operation counts on large-scale, modest-accuracy problems without strong regularity assumptions on the data (Aybat et al., 2013).
  • First-order splitting and operator methods (SCS, SuperSCS, COSMO) are preferred for very large problems or when parallelization is crucial. They benefit substantially from problem-specific data scaling and custom cone projection implementations (O'Donoghue et al., 2013, Sopasakis et al., 2019, Garstka et al., 2019).
  • Interior-point algorithms (alfonso, Clarabel) support both symmetric and nonsymmetric cones, match the best-known theoretical complexity, and seamlessly deliver rigorous certificates. The extension to user-supplied barrier oracles enables direct treatment of conic structures that otherwise admit only cumbersome semidefinite or extended formulations (Papp et al., 2021, Papp et al., 22 Feb 2025, Goulart et al., 21 May 2024).
  • Chordal decomposition techniques for SDPs, and clique merging via clique graphs, yield major speedups for structured optimization—up to nearly 100× observed in large sparse semidefinite test sets (Goulart et al., 21 May 2024, Garstka et al., 2019).

GPU-based acceleration (CuClarabel) complements these trends, as mixed parallel computing and mixed-precision linear solves further reduce time-to-solution for large-scale, heterogeneous conic optimization (Chen et al., 26 Dec 2024).

7. Extensions, Generalizations, and Future Perspectives

Convex conic optimization continues to expand along several fronts:

  • Generalized frameworks such as CGPs unify previously separate problem classes and facilitate entropic and exponential constraints with conic structure (Chandrasekaran et al., 2013).
  • Systematic conification of generalized disjunctive programs enables transformation of convex GDPs into MICP representations solvable by conic solvers, capitalizing on extended formulations and perspective functions (Neira et al., 2021).
  • Advancements in duality theory, infeasibility certification, and initialization (e.g., dual certificates, two-phase methods, and homogeneous self-dual embeddings) ensure both rigor and reliability for challenging or degenerate problems (Papp et al., 22 Feb 2025).
  • The interplay between optimization modeling tools and advanced conic solvers accelerates uptake of advanced methods in application domains, supported by interface frameworks such as Convex.jl, JuMP, and CVXPY (Udell et al., 2014, Goulart et al., 21 May 2024).

Remaining challenges and directions encompass further scalability advances (notably for the hardest polynomial optimization and high-dimensional SDPs), broader support for nonsymmetric and hyperbolic cones, and automated exploitation of problem structure for optimal splitting, decomposition, and hybrid CPU–GPU computation (Chen et al., 26 Dec 2024, Papp et al., 22 Feb 2025).


Convex conic optimization provides a unifying lens for computational convexity, supporting theoretical foundations, algorithmic innovation, and widespread deployment in large-scale, modern applications. The mature synergy between duality theory, barrier methods, and operator splitting underpins its continuing evolution and foundational role in mathematical optimization.