Papers
Topics
Authors
Recent
2000 character limit reached

Krasnoselskii-Mann Iteration

Updated 4 January 2026
  • Krasnoselskii-Mann Iteration is an iterative method that computes fixed points of nonexpansive operators via a weighted blend of the current iterate and its operator evaluation.
  • The algorithm achieves convergence under divergent-series step-size conditions and has been extended to include inertial, stochastic, and accelerated schemes with robust performance.
  • Its practical applications span convex optimization, signal processing, and reinforcement learning, offering reliable solutions even amid perturbations and adaptive updates.

The Krasnoselskii-Mann (KM) iteration is a foundational iterative scheme for finding fixed points of nonexpansive operators in linear and nonlinear settings. It underpins a wide range of algorithms in monotone operator theory, convex optimization, game theory, stochastic approximation, and reinforcement learning. KM-type schemes balance contraction (relaxation) with inertia and are robust to perturbations, variable step sizes, and stochastic noise. Their convergence properties, rate bounds, and generalizations have been the subject of intensive analysis, leading to sharp characterizations in Hilbert, Banach, CAT(0), and even general normed spaces.

1. Definition, Origin, and Operator-Theoretic Formulation

The Krasnoselskii-Mann iteration is defined by

xn+1=(1αn)xn+αnT(xn)x_{n+1} = (1-\alpha_n)x_n + \alpha_n T(x_n)

where T:XXT:X\to X is a nonexpansive operator (TxTyxy\|T x - T y\|\le\|x - y\|), (X,)(X,\|\cdot\|) is typically a Hilbert or Banach space, and {αn}n0[0,1]\{\alpha_n\}_{n\ge0}\subset[0,1] is a sequence of relaxation parameters.

Classically, convergence is guaranteed if n=0αn(1αn)=+\sum_{n=0}^\infty \alpha_n(1-\alpha_n)=+\infty, a condition ensuring the scheme does not stagnate (Cominetti et al., 2012). Equivalently, the KM iteration can be interpreted in terms of averaged operators: TT is θ\theta-averaged if T=(1θ)Id+θRT = (1-\theta)\mathrm{Id} + \theta R for a nonexpansive RR. The iteration generalizes Banach-Picard, Krasnoselskiĭ, and Mann schemes, and subsumes the Halpern iteration as a limiting case (Bot et al., 2022).

For linear TT, and in multi-agent systems and game theory, the KM iteration connects fundamentally to the concept of strict pseudocontractiveness, which underpins necessary and sufficient conditions for convergence (Belgioioso et al., 2018).

2. Convergence Theorems and Rate Analysis

Weak and Strong Convergence

KM iterations converge weakly to a fixed point under nonexpansive TT and the divergent-series step-size condition αn(1αn)=\sum \alpha_n(1-\alpha_n)=\infty (Cominetti et al., 2012). For linear nonexpansive TT, this iteration converges \emph{strongly} (in norm) to the metric projection of x0x_0 onto FixT\operatorname{Fix} T, requiring only the divergent-series condition, without uniform positive lower/upper bounds on αn\alpha_n (Bartz et al., 28 Dec 2025). This generalizes the Baillon–Bruck–Reich theorem and later results by Bauschke–Combettes.

In uniformly convex Banach spaces, asymptotic regularity and convergence rates can be quantified using "proof mining" techniques; for constant αn=α(0,1)\alpha_n=\alpha\in(0,1), the quadratic rate xnTxn=O(1/n)\|x_n-Tx_n\|=O(1/n) is attainable in Hilbert spaces and under suitable regularity conditions (Firmino et al., 16 Jan 2025).

Rate Bounds and Optimality

The sharp universal upper bound for the fixed-point residual is

xnTxndiam(C)i=1nαi(1αi)\|x_n - T x_n\| \le \frac{ \operatorname{diam}(C) }{ \sqrt{ \sum_{i=1}^n \alpha_i(1-\alpha_i) } }

and this rate O(1/n)O(1/\sqrt{n}) is optimal for general normed spaces (Cominetti et al., 2012, Contreras et al., 2021), with Halpern-type iterations attaining O(1/n)O(1/n), which KM cannot generally match. The residual decay rates for various schemes are summarized below.

Scheme Typical Rate Source(s)
Classical KM O(1/n)O(1/\sqrt{n}) (Cominetti et al., 2012, Contreras et al., 2021)
Halpern (linear/firm) O(1/n)O(1/n) (Contreras et al., 2021)
Generalized stochastic Linear & Quadratic (O(1/n)O(1/n), O(1/n2)O(1/n^2)) (Pischke et al., 2024)
Fast/Nesterov-KM o(1/k)o(1/k) (momentum, accelerated) (Bot et al., 2022, He et al., 28 Oct 2025)
AdaGrad-regret-KM O(1/T)O(1/\sqrt{T}), data-adaptive (Kwon, 25 Sep 2025)

3. Extensions: Inertia, Perturbation, and Generalized Schemes

Inertial Krasnoselskii-Mann (IKM)

IKM iterations inject momentum: wk=xk+βk(xkxk1),xk+1=(1αk)wk+αkT(wk)w_k = x_k + \beta_k(x_k - x_{k-1}), \quad x_{k+1} = (1-\alpha_k) w_k + \alpha_k T(w_k) with βk\beta_k controlling inertia (Cui et al., 2019, Maulén et al., 2022). Provided parameter sequences and error terms are summable, weak (and under quasi-contractive maps, strong/linear) convergence is established, along with O(1/k)O(1/k) nonasymptotic bounds on the best residual (Cui et al., 2019).

The practical advantage is acceleration, observable in primal-dual splitting and multi-operator monotone inclusions. Inertia allows faster empirical convergence at modest risk of divergence, necessitating careful parameter selection (Maulén et al., 2022).

Stochastic and Inexact Iterations

Noise-resilient versions (with martingale-difference or arbitrary perturbations) admit almost sure convergence and explicit nonasymptotic residual bounds. Under bounded variance or summable errors, rates mirror deterministic O(1/n)O(1/\sqrt{n}) decay, with minor log or step-size corrections (Bravo et al., 2017, Bravo et al., 2022, Sababe et al., 2 Jun 2025).

The general proof architecture for stochastic variants relies on the Robbins–Siegmund supermartingale lemma, and Fejér-type monotonicity, which extends to adaptive Bregman geometries and heavy-tailed noise models (Sababe et al., 2 Jun 2025).

4. Advanced Schemes and Algorithmic Accelerations

Nesterov-Type and Adaptive-Momentum Variants

Momentum-accelerated KM algorithms, such as Fast KM and TKMA, utilize Nesterov's extrapolation or adaptive local geometry, blending information from T2T^2 (Picard) and momentum steps, with analytically derived or geometrically motivated momentum parameters (Bot et al., 2022, He et al., 28 Oct 2025). The resulting schemes achieve o(1/k)o(1/k) or o(1/k1/2)o(1/k^{1/2}) rates on iterate differences and often outperform classical and Halpern algorithms in image denoising and matrix completion applications (He et al., 28 Oct 2025).

Tikhonov Regularization and Forward-Backward Splitting

Tikhonov regularization augments KM with shrinking steps: xn+1=(1λn)βnxn+λnTn(βnxn)x_{n+1} = (1-\lambda_n) \beta_n x_n + \lambda_n T_n(\beta_n x_n) where βn1\beta_n \to 1 and λn\lambda_n control regularization and relaxation, respectively (Bot et al., 2019). This yields strong convergence to the minimal-norm solution for countable families of operators and in monotone inclusion settings, especially when coupled with variable step sizes (forward-backward algorithms).

The framework robustly accommodates errors and variable steps, directly translating into accelerated splits for convex optimization and signal processing (Bot et al., 2019).

5. Geometric, Nonlinear, and Game-Theoretic Generalizations

Nonlinear (CAT(0), Hyperbolic) KM Iterations

KM extends to metric and geodesic spaces (CAT(0)), traditionally formulated as: xn+1=(1αn+1)xnαn+1T(xn)x_{n+1} = (1-\alpha_{n+1}) x_n \oplus \alpha_{n+1} T(x_n) where \oplus denotes the geodesic convex combination (Foglia et al., 29 Oct 2025). Asymptotic regularity is preserved, and the same O(1/n)O(1/\sqrt{n}) rate applies. Convergence to a fixed point (in the sense of Δ\Delta-convergence) is proved under mild control on step sizes.

Halpern iteration in metric settings further accelerates rates (O(1/n)O(1/n)) and motivates hyperbolic variants for deep learning optimizers (Foglia et al., 29 Oct 2025).

Consensus, Equilibrium, and Relative KM Iteration in Games

In multi-agent consensus problems, KM iteration finds equilibria even when network topology is only partially known. Convergence is guaranteed iff the underlying operator is strictly pseudocontractive, connectable to spectral and LMI criteria (Belgioioso et al., 2018).

In stochastic mean-payoff and entropy games, "relative" KM schemes leverage normalized operators under Hilbert seminorms, exploiting the additive homogeneity of Shapley operators to achieve O(logϵ)O(|\log \epsilon|) complexity for ϵ\epsilon-approximation, significantly improving upon prior bounds (Akian et al., 2023).

6. Practical Implementations and Application Domains

KM-type iterations and their variants (stochastic, inertial, adaptive, Tikhonov-regularized) underpin algorithms in:

Empirical comparisons show that momentum-enhanced and adaptive-KM variants consistently outperform classical schemes in both computational speed and convergence rate, but can exhibit oscillatory or spiraling behavior typical of momentum methods (Bot et al., 2022, He et al., 28 Oct 2025).

7. Parameter Tuning, Rate Explicitness, and Theoretical Implications

Quantitative bounds on asymptotic regularity and convergence rates are now explicit due to advances in proof mining and optimal transport analysis (Firmino et al., 16 Jan 2025). This enables detailed complexity planning for practical implementations, for instance, explicit oracle complexity O(ε5)O(\varepsilon^{-5}) for minibatch KM in stochastic environments (Pischke et al., 2024). Parameter dependencies on the relaxation and inertia sequences, convexity moduli, and problem geometry guide optimal algorithm design.

KM-type schemes are robust to perturbations, approximation, and stochastic deviations, provided error terms are appropriately controlled (summable or diminishing), and inertia/momentum parameters are carefully chosen to avoid instability (Maulén et al., 2022, Cui et al., 2019).

References

Summary Table: Scheme Variants and Convergence Properties

Iteration Variant Rate & Convergence Noise/Stability Parameter Control
Classical KM O(1/n)O(1/\sqrt{n}), weak (linear: strong) Robust to summable errors αn(1αn)=\sum \alpha_n(1-\alpha_n)=\infty
Inertial KM O(1/k)O(1/k)o(1/k)o(1/k) best residual Sensitive to inertia choice Bounded/increasing βk\beta_k
Tikhonov-KM Strong convergence, min-norm Variable step size, small Tikhonov βn1\beta_n\to1, decay λn\lambda_n
Fast, Adaptive KM o(1/k)o(1/k), o(1/k)o(1/\sqrt{k}) Mildly oscillatory Momentum schedule, geometric
Stochastic/Bregman SKM O(ANp)O(A_N^{-p}) residual averages Martingale/noise trimming αk2<\sum \alpha_k^2 < \infty
Proof-mined Generalized Explicit Φ(k)\Phi(k) rate, quadratic possible Error modulus imposed Fejér, convexity modulus

The Krasnoselskii-Mann iteration remains a central tool in nonlinear analysis, monotone operator theory, and optimization. Its generalizations, inertial and regularized variants, stochastic extensions, and precise rate theory constitute a mature and versatile algorithmic arsenal with rigorous theoretical guarantees and broad practical impact across mathematical and computational sciences.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (17)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Krasnoselskii-Mann Iteration.