Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 86 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 43 tok/s
GPT-5 High 37 tok/s Pro
GPT-4o 98 tok/s
GPT OSS 120B 466 tok/s Pro
Kimi K2 225 tok/s Pro
2000 character limit reached

Generalized Newton Method

Updated 18 August 2025
  • Generalized Newton Method is a class of algorithms that extend classical Newton’s method using graphical derivatives to linearize nonsmooth mappings.
  • It leverages metric regularity and coderivative conditions to ensure robust, often superlinear, local and global convergence in complex variational problems.
  • The method broadens the applicability beyond semismooth approaches, enabling practical solutions for nonsmooth, set-valued equations and variational inequalities.

The Generalized Newton Method comprises a class of algorithms that extend the classical Newton’s method to broad families of nonsmooth and set-valued equations, including generalized equations and inclusions defined by variational analysis. These methods are designed for problems where the underlying mapping may be nonsmooth, non-Fréchet differentiable, or even set-valued, and in which classical smooth calculus does not suffice. The principal innovation is to employ generalized differential tools—such as graphical derivatives, coderivatives, or related generalized Jacobians—to construct Newton-type iterations with strong local (often superlinear) and, in favorable cases, global convergence guarantees, frequently under metric regularity hypotheses.

1. Graphical Derivative-Based Newton Algorithm

The core problem is to solve nonsmooth equations H(x)=0H(x) = 0 for continuous, but potentially nonsmooth, mappings H:RnRnH: \mathbb{R}^n \to \mathbb{R}^n. The fundamental step in the generalized Newton method based on graphical derivatives is to define the search direction dkd^k at each iterate xkx^k as a solution to the inclusion

H(xk)DH(xk)(dk),-H(x^k) \in D H(x^k) (d^k),

where DH(xk)D H(x^k) denotes the graphical derivative of HH at xkx^k. In the smooth case, this reduces to solving H(xk)=H(xk)dk-H(x^k) = H'(x^k) d^k, the classical Newton step. For nonsmooth HH, the graphical derivative DH(x)DH(x)—which captures the directional limiting behavior of the mapping—allows for the linearization of HH in directions where classical derivatives are undefined.

The numerical iteration proceeds as:

xk+1=xk+dkx^{k+1} = x^k + d^k

with dkd^k determined as above. Well-posedness (existence of dkd^k) and local convergence require HH to be metrically regular at a solution—this allows the graphical derivative inclusion to be solvable and the iteration to progress.

2. Role of Metric Regularity and Uniform Error Estimates

Metric regularity serves as the cornerstone for the convergence analysis of the graphical-derivative-based Newton scheme. A mapping HH is metrically regular at xx^* for $0$ with modulus μ>0\mu > 0 if, locally,

dist(x,H1(0))μH(x)for x near x.\text{dist}(x, H^{-1}(0)) \leq \mu \| H(x) \| \quad \text{for } x \text{ near } x^*.

This property is characterized via coderivatives, specifically, by the invertibility condition of the graphical derivative at xx^*. The metric regularity hypothesis ensures not only well-posedness of the Newton subproblem but also enables the derivation of crucial uniform error bounds.

A central result is that, under metric regularity with modulus μ\mu and for iterates sufficiently close to the solution xx^*,

xk+1xμ1μxkxk1.\| x^{k+1} - x^* \| \leq \frac{\mu}{1 - \mu} \| x^k - x^{k-1} \|.

This error estimate, analogous to those in certain proximal-point schemes [Artacho and Goeffroy], quantifies the contraction and robustness of the iteration—even permitting inexact computation of the direction dkd^k if the errors are controlled relative to μ\mu.

3. Variational Analysis and Generalized Differentiation in Algorithm Construction

The development and analysis of the generalized Newton method leverage advanced tools from variational analysis:

  • Graphical derivatives (DHDH): Capture the tangent cone to the graph of HH at a point and generalize the Jacobian for nonsmooth mappings.
  • Coderivatives: Provide the adjoint to the graphical derivative, essential for analyzing metric regularity and error bounds.
  • Generalized Jacobians and limiting subdifferentials: Supply the calculus for constructing and analyzing the iteration in the absence of smoothness.

Specifically, the coderivative-based characterization of metric regularity (involving surjectivity-type or invertibility-type conditions) underlies both the existence theory for the Newton subproblem and the convergence rate guarantees.

4. Comparison with Semismooth and BB-Differentiable Newton Methods

Traditional Newton-type schemes for nonsmooth problems—such as semismooth and BB-differentiable Newton methods—typically restrict their application to Lipschitzian equations satisfying semismoothness or BB-differentiability. These rely on the generalized Jacobian in the sense of Clarke or on some special directional differentiability condition.

The graphical derivative-based Newton method admits cases beyond the semismooth framework by not imposing any semismoothness assumptions on HH. In situations where semismoothness fails (i.e., the nonlinearity exhibits nondirectional or irregular behavior), the method remains applicable, provided metric regularity holds. Thus, the approach properly contains the semismooth Newton method as a special case and extends applicability and robustness.

5. Global and Superlinear (Kantorovich-Type) Convergence

Beyond local behavior, the graphical derivative-based generalized Newton method admits a Kantorovich-type global convergence theorem: if the mapping HH is metrically regular and certain further continuity and regularity (assumptions (H1) and (H2) in the original presentation) are met, then the iteration converges globally from initial guesses sufficiently close to a solution with superlinear rate.

Superlinear convergence follows from the fact that, under these hypotheses, the error between iterates contracts faster than linearly:

xk+1x=o(xkx),\| x^{k+1} - x^* \| = o( \| x^k - x^* \| ),

assuming the graphical derivative mapping retains suitable regularity in a neighborhood.

6. Practical Implications and Robustness to Nonsmoothness

The method allows for practical robustness to the lack of classical differentiability:

  • No requirement of semismoothness: Graphical derivatives support mappings exhibiting genuine nonsmoothness.
  • Tolerance to inexact computation: Convergence is retained even if the Newton direction is computed approximately, provided the error is small relative to the metric regularity modulus.
  • Applicability to nonsmooth equation classes: Inclusion problems, nonsmooth variational equalities, and non-Lipschitz mappings are encompassed if metric regularity and associated coderivative estimates hold.

This broadens the relevance of Newton-type methods to settings previously inaccessible to standard or semismooth Newton methods, with significant implications for variational inequalities, nonsmooth optimization, and operational models involving set-valued mappings.

7. Connections to Proximal Frameworks and Broader Variational Methodologies

There is a strong parallel between the graphical derivative-based Newton method and inexact proximal-point algorithms analyzed under metric regularity [Artacho and Goeffroy]. Both frameworks rely upon error estimates tied to metric regularity and employ variational analytic constructs (typically coderivatives) to ensure robustness in the face of nonsmoothness and computational inexactness. This conceptual alignment suggests that the convergence guarantees, error bounds, and regularity results obtained for the Newton-type method have immediate implications for proximal-type methods, and vice versa, reinforcing the centrality of metric regularity in modern variational numerical analysis.