Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
GPT-5.1
GPT-5.1 130 tok/s
Gemini 3.0 Pro 29 tok/s Pro
Gemini 2.5 Flash 145 tok/s Pro
Kimi K2 191 tok/s Pro
Claude Sonnet 4.5 34 tok/s Pro
2000 character limit reached

Second-Order Quasi-Newton Algorithm

Updated 16 November 2025
  • Second-order quasi-Newton algorithm is a numerical optimization method that approximates the inverse Hessian using secant conditions to capture curvature for rapid convergence.
  • The algorithm employs iterative updates such as BFGS, DFP, and SR1, effectively balancing computational cost with improved stability, even under noisy gradients.
  • Modern variants extend this method to distributed and high-dimensional problems by incorporating limited-memory techniques and stochastic regularization to ensure robust performance.

A second-order quasi-Newton algorithm is a numerical optimization strategy that seeks to approximate the Newton-Raphson direction for solving unconstrained or constrained nonlinear problems. Unlike first-order methods, which leverage only gradient (first derivative) information, quasi-Newton approaches build an explicit estimate of the inverse Hessian (second derivative) or its action, typically through low-rank iterative updates based exclusively on observed iterates and gradients. These algorithms aim to accelerate convergence by capturing curvature, often using little memory and low per-iteration computational cost. The foundational update mechanism is rooted in the secant condition, which stipulates Bk+1sk=ykB_{k+1} s_k = y_k for appropriate choices of sks_k and yky_k, with Bk+1B_{k+1} acting as a surrogate Hessian. Modern advances address stochasticity, high-dimensionality, nonconvexity, parallelism, and non-Euclidean geometric structure.

1. Core Principles and Update Mechanisms

Quasi-Newton algorithms begin with the goal of achieving rapid convergence akin to Newton's method, while circumventing the prohibitive costs associated with computing or inverting the true Hessian. The key technical device is the secant condition, which for sk=xk+1xks_k = x_{k+1} - x_k and yk=f(xk+1)f(xk)y_k = \nabla f(x_{k+1}) - \nabla f(x_k) imposes

Bk+1sk=yk,B_{k+1} s_k = y_k,

where Bk+1B_{k+1} is the next iterate's Hessian approximation. Classical update schemes include BFGS, DFP, and SR1, each built to preserve symmetry, positive definiteness, and invariance properties where possible.

For large-scale problems, the limited-memory variant (L-BFGS) forms the inverse Hessian implicitly, storing only mm recent pairs (sk,yk)(s_k, y_k) and applying the approximation via the two-loop recursion. This enables scalability to n106n\gg10^6 with O(mn)O(mn) storage and computational complexity per iteration.

Extension to stochastic regimes necessitates modifications to the secant construction and aggregation rules. For example, SpiderSQN (Zhang et al., 2020) uses variance-reduced stochastic gradients in tandem with damped L-BFGS updates, ensuring sTy>0s^T y > 0 by combining gradient differences and regularization schemes, crucial for maintaining stability in noisy environments.

More sophisticated quasi-Newton variants leverage multi-secant conditions, simultaneously enforcing BS=YB S = Y for secant matrices SS and YY constructed from multiple steps, and restrict updates to symmetric positive-definite (SPD) perturbations to ensure descent directions (Lee et al., 9 Apr 2025).

2. Scaling, Initialization, and Phenomena Handling

Initialization and scaling of the Hessian approximation are critical, particularly in regimes characterized by "vanishing" or "exploding" gradients. The adaQN algorithm (Keskar et al., 2015) addresses this by adopting an Adagrad-style diagonal scaling for Hk(0)H_k^{(0)}, specifically

[Hk(0)]ii=1j=0k[gj]i2+ϵ,[H_k^{(0)}]_{ii} = \frac{1}{\sqrt{ \sum_{j=0}^k [g_j]_i^2 + \epsilon } },

replacing naive scalar initialization that leads to poor numerical conditioning in recurrent neural networks.

This diagonal preconditioning stabilizes updates prior to accumulating meaningful curvature statistics. AdaQN further judiciously accepts new curvature pairs (s,y)(s,y) only when sTy>ϵs2s^Ty > \epsilon \|s\|^2, rejecting updates that could degrade the predictive quality or induce instability. Aggregated iterates and Fisher information matrix (FIM)-based gradient products are used to robustly estimate yy in the presence of stochastic gradients.

3. Extensions to Structured, Distributed, and High-Dimensional Problems

Distributed second-order quasi-Newton schemes have been developed to address the communication bottleneck and lack of global curvature modeling in massive data settings. In methods such as DPLBFGS and distributed SpaRSA (Lee et al., 2019, Lee et al., 2018), the model

mk(d)=f(xk)+f(xk)Td+12dTBkd+Ψ(xk+d)Ψ(xk),m_k(d) = f(x_k) + \nabla f(x_k)^T d + \frac{1}{2} d^T B_k d + \Psi(x_k + d) - \Psi(x_k),

is solved in a coordinated fashion across KK processing nodes. Here, BkB_k is constructed using L-BFGS with global curvature pairs, and subproblems are solved via distributed proximal-gradient methods, incurring only O(d)O(d) communication per main iteration. This circumvents stagnation typical of block-diagonal-only Hessian approximations.

Decentralized quasi-Newton updates on networks (Eisen et al., 2016) define local secant conditions and regularized BFGS updates over neighborhood variables, achieving affine invariance and robustness in synchronous and asynchronous multi-agent systems.

4. Algorithmic Complexity, Memory, and Convergence Properties

Second-order quasi-Newton algorithms achieve O(n)O(n) per-step computational complexity in well-designed limited-memory implementations, as in adaQN (Keskar et al., 2015) and SpiderSQN (Zhang et al., 2020). Complexity for constructing and applying mm low-rank updates scales as O(mn)O(m n), with amortized curvature computation of O(n)O(n) per iteration when the memory window and batch size are constant.

Global convergence rates have been established for both convex and nonconvex objectives. For example, SpiderSQN achieves SFO complexity of O(n+n1/2ϵ2)O(n + n^{1/2} \epsilon^{-2}) for finding ϵ\epsilon-stationary points under smoothness and bounded Hessian conditions (Zhang et al., 2020). In distributed convex ERM, linear communication-efficient rates are proved: F(xt+1)F(1ρ)(F(xt)F),F(x^{t+1}) - F^* \leq (1 - \rho)\left( F(x^t) - F^* \right), for some ρ(0,1)\rho \in (0,1), even under block-separable nonsmooth regularization (Lee et al., 2019, Lee et al., 2018).

In strongly convex quadratic settings, memory-two quasi-Newton updates combined with Krylov subspace projections can guarantee finite termination independent of line search precision (Ansari-Önnestam et al., 3 Jul 2024).

Higher-order and saddle-point extensions are also realized via quasi-Newton updates approximating powers or squares of indefinite Hessians, restoring positive-definiteness in nonconvex or min-max problems (Liu et al., 2021).

5. Advanced Variants and Robustness: Divergence, Regularization, and Error Analysis

Recent advances generalize quasi-Newton updates beyond KL-divergence-based formulations. Bregman extensions (Kanamori et al., 2010) incorporate a broader family of divergence functions, producing self-scaling update rules: Bk+1=θkBBFGS+(1θk)ykykTskTyk,B_{k+1} = \theta_k B^{\textrm{BFGS}} + (1-\theta_k)\frac{y_k y_k^T}{s_k^T y_k}, with θk\theta_k determined via consistency conditions on determinants. Robustness analysis using influence functions demonstrates that only the classical BFGS update yields bounded sensitivity under line-search errors; dual and variant updates can have unbounded error amplification.

Stochastic line-search quasi-Newton frameworks (Wills et al., 2019) further regularize step-size selection via noise-aware Armijo rules, guaranteeing convergence in expectation even under nonvanishing gradient noise, by monitoring sufficient decrease adjusted for noise variance.

Higher-order tensor and Riemannian extensions accommodate manifold domains and non-Euclidean ground metrics. Adaptive cubic-regularization quasi-Newton techniques achieve O(ϵg3/2)O(\epsilon_g^{-3/2}) first-order iteration complexity even when gradients and Hessians are accessed only via finite-differences (Louzeiro et al., 19 Feb 2024).

6. Practical Implementation and Empirical Results

Limited-memory implementations utilize two-loop recursions for Hessian–vector products to eliminate the need for explicit matrix storage. In very large deep learning applications, stochastic quasi-Newton methods employing Gauss–Newton approximation and SVRG-style variance reduction (e.g., SQGN (Thiele et al., 2020)) enable practical second-order optimization in frameworks such as TensorFlow, demonstrating superior test accuracy over Adam/SGD at moderate computational overhead.

Empirical evaluations across distributed logistic regression, ERM, SVM, and nonconvex neural network training confirm that properly regularized and curvature-aware quasi-Newton variants consistently outperform first-order methods when fast local convergence and robustness to ill-conditioning are required. Multi-secant methods further reduce iteration counts on convex tasks, but require algorithmic care (PSD symmetrization and diagonal boosting) to maintain stability outside the quadratic regime (Lee et al., 9 Apr 2025).

7. Contemporary Directions and Open Problems

Active research focuses on algorithmic developments that strike an optimal trade-off between curvature modeling and computational or communication overhead. This includes variance reduction in stochastic settings, asynchronous and decentralized operation for federated systems, sophisticated regularization and safeguarding for robust operation under uncertainty, and advanced multi-secant and higher-order geometric schemes.

Challenges remain in extending guaranteed convergence to highly nonconvex objectives, in controlling error propagation under limited memory and noisy gradients, and fully automating hyper-parameter selection, scaling initialization, and curvature memory management for large-scale heterogeneous architectures.

Ongoing work in the refinement of update formulae, theoretical characterization of robustness, and empirical validation in modern machine learning continues to expand the applicability of second-order quasi-Newton algorithms, making them central tools in both theory and practice.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Second-Order Quasi-Newton Algorithm.