Invariant Krylov Subspace
- Invariant Krylov subspace is the space spanned by successive applications of an operator to a vector, ceasing to yield new directions once invariance is reached.
- It underlies iterative solvers like GMRES and CG, enabling efficient solutions of linear systems, eigenvalue problems, and model reduction in complex computations.
- Extensions such as block and rational Krylov subspaces enhance shift invariance and structure preservation, critical for quantum algorithms and large-scale optimization.
An invariant Krylov subspace is a central concept in numerical linear algebra and computational mathematics, capturing the subspace structure generated by repeated action of an operator (typically a matrix) on an initial vector or set of vectors. Krylov subspaces are pivotal for iterative methods aimed at solving linear systems, eigenvalue problems, and for model reduction in high-dimensional settings. Their invariance properties are foundational both in classical and modern computational frameworks, including applications in large-scale optimization, eigensolvers, quantum algorithms, and scientific computing.
1. Definition of Invariant Krylov Subspace
A Krylov subspace of order associated with matrix and vector is defined as
For block or rational variants, the initial generator may be a matrix and the iterative actions may include rational functions or blocks.
The subspace is termed invariant under if for some . In practice, “invariant Krylov subspace” often refers to the smallest for which this property holds: at this point the sequence no longer produces new directions and the Krylov subspace encodes all information about acting on (1811.09025).
Krylov subspaces can also be generalized to rational forms,
which provide enhanced invariance properties for shifted systems and model reduction (Daas et al., 30 Jun 2025).
2. Construction and Properties
The standard Arnoldi iteration constructs an orthonormal basis for the Krylov subspace. The procedure
- Starts with (where is an initial residual),
- At step forms , orthogonalizes against previous , and normalizes,
- Iterates until breakdown occurs (), which signals that , i.e., the subspace is invariant (1811.09025).
In specialized settings, the basis may be unnormalized (Lanczos triples: with ) (Forsgren et al., 2014) or built to satisfy additional structure (e.g., -orthogonal, symplectic) (Benner et al., 2022, Koskela, 2017).
For methods such as Krylov subspace descent for deep learning, basis construction involves preconditioning and explicit orthogonalization steps: where is a diagonal preconditioner and the gradient (Vinyals et al., 2011).
An important property: for a matrix , Krylov subspaces generated by (or shifted versions for any shift ) from the same initial vector are shift-invariant, i.e., (Burke, 2022, Daas et al., 30 Jun 2025).
3. Role in Iterative Solvers and Optimization
Krylov subspaces are the foundation for iterative methods including Conjugate Gradient (CG), GMRES, MINRES, and their variants (1811.09025, Forsgren et al., 2014). These approaches iteratively seek solutions in growing Krylov subspaces and exploit their invariance:
- For symmetric positive definite, CG generates -conjugate directions in .
- GMRES, through the Arnoldi process, constructs a basis for and seeks the residual minimizer in this subspace (1811.09025).
In high-dimensional optimization, such as cubic-regularized Newton updates, Krylov subspaces built from the Hessian and gradient allow for low-dimensional approximations of the Newton step while maintaining dimension-independent convergence rates (Jiang et al., 5 Jan 2024). The subspace is invariant up to order and contains directions capturing the essential spectrum of the Hessian.
Enlarged Krylov subspaces—constructed by iteratively applying to multiple domain-decomposed blocks per iteration—capture a larger portion of the invariant subspace per iteration, leading to faster convergence and reduced communication overhead in parallel/distributed settings (Moufawad, 2018).
4. Invariance in Block, Rational, and Structure-Preserving Extensions
- Block Krylov subspaces: For and block ,
Block Krylov subspaces support subspace expansions optimized to converge to target invariant subspaces with minimal principal angles (Zuccalli et al., 21 Nov 2024).
- Rational Krylov subspaces: Allow inclusion of shifted/inverse directions providing enhanced approximation of spectra near poles and shift-invariant properties critical for sequences of shifted systems (Daas et al., 30 Jun 2025).
- Structure-preserving Krylov subspaces: In Hamiltonian systems, basis construction must be symplectic (e.g., -orthogonal) to ensure preservation of invariants such as energy (Koskela, 2017, Benner et al., 2022). For skew-Hamiltonian matrices, Krylov subspaces are automatically isotropic and can be constructed to yield any Lagrangian subspace (Saltenberger et al., 2019).
5. Augmentation, Recycling, and Adaptivity
Invariant subspace information can be amplified or recycled across multiple solves:
- Augmented Krylov methods: These enhance by an additional fixed or recycled subspace that approximates invariant directions—beneficial for accelerating convergence especially for slow modes associated with near-invariant subspaces (Soodhalter, 2021).
- Recycling for evolving systems: In applications such as PDE-constrained optimization with changing meshes or matrices, previously computed invariant subspaces are mapped to new algebraic contexts via interpolation and further refined by warm-start Krylov–Schur cycles (Bolten et al., 2020).
- Pole selection/adaptive expansion: In rational Krylov settings for sequences of shifted systems, adaptively choosing the next pole to match the shift with the largest residual error ensures that the new subspace captures the most difficult direction, preserving the property that the solution to certain shifted systems is exactly included after the pole is matched (Daas et al., 30 Jun 2025).
6. Connections to Quantum Algorithms and Model Reduction
In quantum simulation and quantum eigensolver algorithms:
- Quantum Krylov subspace diagonalization (QKSD): Approximates eigenstates in an invariant subspace spanned by , often with structured (e.g., Hankel) representations. Advanced techniques such as quantum signal processing are employed to prepare these states and extract molecular gradients and observables with measurement cost scaling reduced from to (Oumarou et al., 9 Jan 2025).
- Measurement-efficient Gaussian-power Krylov bases: Exploit the invariance of filtered Krylov subspaces to suppress statistical noise while retaining convergence properties (Zhang et al., 2023).
In model reduction (e.g., Riccati and Lyapunov equations), invariant (Krylov or low-rank) subspaces yield numerical schemes for operator equations with reduced computational and storage costs while maintaining theoretical error bounds (Behr et al., 2019).
7. Optimality, Error Bounds, and Computable Algorithms
Optimal iterative expansion of subspaces to approximate invariant subspaces (e.g., those associated with dominant eigenvalues) may proceed via expansions of the form , where is chosen to minimize principal angles with the target invariant subspace. Theoretical results yield geometric convergence under spectral gap assumptions and lower bounds using Chebyshev polynomial filters (Zuccalli et al., 21 Nov 2024). Practically, computable versions employ variants of Rayleigh–Ritz or refined projection methods for correction.
Bounds on residuals and error are often directly linked to the invariance property:
- If the Krylov subspace becomes exactly invariant, residuals vanish and iteration stops (1811.09025).
- In inner-product free iterative methods, the lack of orthogonality is controlled via condition numbers of well-defined factors, preserving approximation properties of the invariant subspace (Brown et al., 8 Sep 2024).
In summary, invariant Krylov subspaces encapsulate essential algebraic, geometric, and spectral properties necessary for efficient iterative methods in scientific computing, optimization, and quantum simulation. Their construction, structure, and adaptability underpin the design of efficient, robust, and theoretically justified solvers for high-dimensional and evolving problems across classical and quantum computational paradigms.